专利摘要:
A mobile terminal (100, 200) having a side display unit comprises a body having a front surface, side surfaces and a back surface, a display unit (151, 251) having a first region (261) on the front surface and a second region (262) on the lateral surface extending from the first region, a touch detection unit detecting a touch input applied to the display unit, and a control member (180) for detecting a minus one touch input applied to one or more points on the second region in a state where screen information (300) is provided to the first region, and to process the at least one touch input as a storage command from at least a portion of the screen information delivered to the first region when the at least one touch input corresponds to a predetermined touch type.
公开号:FR3019665A1
申请号:FR1552712
申请日:2015-03-31
公开日:2015-10-09
发明作者:Daemyeong Park
申请人:LG Electronics Inc;
IPC主号:
专利说明:

[0001] MOBILE TERMINAL AND MOBILE TERMINAL CONTROL METHOD The present disclosure relates to a mobile terminal, and particularly to a mobile terminal comprising a side display unit. The terminals can be generally classified as mobile / portable terminals or as immobile terminals depending on their mobility.
[0002] Mobile terminals can also be classified as portable terminals or as vehicle-mounted terminals depending on whether a user can directly carry the terminal or not. Mobile terminals have become more and more functional. Examples of such functions include data communications, voice communications, image and video acquisition with a camera, audio recording, playback of music files with a speaker system, and audio playback. displaying images and videos on a display. Some mobile terminals include additional features for playing games, while other terminals are configured as media players.
[0003] More recently, mobile terminals have been configured to receive broadcast and multicast signals for viewing content such as videos and television programs. Efforts are being made to support and increase the functionality of mobile devices. Such efforts include software and hardware enhancements as well as changes and improvements to the structural components. For example, a user interface is provided that allows a user to conveniently and conveniently search for a feature and select it. Since the mobile terminal is considered a personal good expressing the personality of its user, various designs are needed. The designs include modifications and structural improvements to enable a user to use the mobile terminal more conveniently. In the context of these structural modifications and improvements, a lateral display unit and a user interface using it may be envisaged. One aspect of the detailed description provides a mobile terminal capable of receiving an input from a user using a side display unit.
[0004] Another aspect of the detailed description provides a mobile terminal capable of performing a specific function in correspondence with a user input on a side display unit. Another aspect of the detailed description provides a mobile terminal capable of providing an acquisition function corresponding to a gesture of a user with respect to a side display unit. To realize these and other advantages, in accordance with the purpose of this memo, as embodied and broadly described herein, there is provided a mobile terminal comprising: a body having a front surface, surfaces; side and a back surface; a display unit having a first region disposed on the front surface, and a second region extending from the first region, the second region being disposed on the side surface; a touch sensor unit configured to detect a touch input applied to the display unit; and a controller configured to: detect, through the touch sensor unit, at least one touch input applied to one or more points on the second region in a state in which screen information has been delivered to the first region, and processing the at least one touch input applied to one or more points on the second region as a storage command for storing at least a portion of the screen information delivered to the first region when the at least one touch input applied to the one or more points on the second region corresponds to a predetermined type of touch In an embodiment, if a first touch input applied to a first point on the second region and a second touch input applied to a second point on the second region are detected, the controller is configured to adjust at least a portion screen information delivered to the first region as a storage region. In one embodiment, the storage region comprising the at least a portion of the screen information is determined based on at least one of a position of the first point and a position of the second point. , and a distance between the first point and the second point. In one embodiment, the storage region is positioned on at least a portion of the first region horizontally corresponding to the first point and the second point on the second region. The storage region has a vertical length corresponding to the distance between the first point and the second point. In one embodiment, if at least one slip touch extending from at least one of the first touch input and the second touch input meets a preset condition, after setting the storage region, the controller is configured to store screen information included in the storage region. In one embodiment, the controller is configured to store the screen information included in the storage region in response to an input of said at least one slip key starting from one of the first point. and the second point, to another of these. In one embodiment, the controller is configured to adjust at least a portion of the first region as a storage region based on a first touch type of the predetermined touch type of touch applied to the first region. second region. The controller is configured to output to the storage region a guide image indicating that the storage region has been set. The guide image is superimposed on the screen information included in the storage region, so that the screen information included in the storage region is recognizable to a user. In one embodiment, the guide image has an output size that is changed based on a second touch type different from the first touch type, from the predetermined touch touch type applied to the second region. The output size of the guide image is progressively reduced based on the second type of touch. In a state in which the output size of the guide image has become smaller than a reference size according to the second type of touch, the storage region is acquired. In one embodiment, even though the output size of the guide image is changed according to the second touch type, an output of a boundary image indicating the storage region is maintained on edges of the region of storage.
[0005] In one embodiment, the guide image has a size corresponding to a size of the storage region. In a state in which the guide image has been delivered, if a third type of touch different from the first type of touch is applied to the second region, the guide image is moved according to the third type of touch. The storage region is reset to a region in which the guide image has been moved. In one embodiment, in a state in which the guide image has been delivered to the first region, if a sliding feel is applied to the guide image, the controller is configured to move the image in a direction of entry of the sliding touch. The controller is configured to reset a region in which the guide image has been moved as a storage region. In one embodiment, in a state in which the guide image has been delivered to the first region, if a touch input is detected on a portion of the first region in which the guide image has not been delivered , the controller is configured to complete delivery of the guide image. The controller is configured to process the detected touch input on the portion of the first region in which the guide image has not been output as a control command to cancel the storage command.
[0006] In one embodiment, in a state in which the guide image has been delivered to the first region, whether a narrowing pinch touch input or a widening nip touch input applied in the second region is detected, the The controller is configured to change a vertical size of the storage region based on the detected narrowing pinch feel or the detected widening nip. The controller is configured to change the output size of the guide image to match the changed vertical size. In one embodiment, the display unit further includes a third region disposed on a side surface opposite the second region.
[0007] In a state in which the guide image has been delivered to the first region, if a narrowing nip or a widening nip applied to the third region is detected, the controller is configured to change. a horizontal size of the storage region based on the detected narrowing pinch touch input or the widening nip touch input detected. The controller is configured to change the output size of the guide image to match the changed horizontal size.
[0008] In one embodiment, if the at least a portion of the screen information is text, the controller is configured to output a graphical user interface (GUI) screen relating to a selection of a storage type. so that a storage format with respect to said at least a portion of the screen information can be selected by a user.
[0009] Said at least a portion of the screen information is stored as at least one of an image file format and a text file format based on a selection of a user . In one embodiment, the controller is configured to store the at least a portion of the screen information in a memory in response to the storage command. The controller is configured to search for an application to perform a function using the at least a portion of the screen information, and to deliver an icon relating to the desired application. In one embodiment, the predetermined tactile touch type includes a first touch input applied to the second region, and a second touch input subsequent to the first touch. If the first touch input is detected, the controller is configured to execute a storage mode for storing said at least a portion of the screen information. If the second touch input following the first touch is detected, the controller is configured to store the at least a portion of the screen information. If the first touch input applied to the second region is released, before detecting the second touch input after detecting the first touch input, the controller is configured to complete the storage mode.
[0010] In one embodiment, in a case in which the screen information supplied to the display unit corresponds to a preset application, even if the predetermined touch type of the touch applied to the second region is detected, the control is configured not to process the detected touch input as a storage command to store at least a portion of the screen information delivered in the first region. The controller is configured to process the detected touch input as a control command to perform a function relating to the preset application.
[0011] To achieve these and other advantages, in accordance with the subject matter of this specification, as embodied and broadly described herein, there is also provided a method of controlling a mobile terminal comprising a front surface, a side surface and a back surface, the method comprising: providing screen information in a first region disposed on the front face; detecting a touch on a second region extending from the first region and disposed on the side face, in a state in which the screen information has been delivered to the first region; the detected touch processing, as a control command for storing at least a portion of the screen information delivered in the first region; and storing said at least a portion of the screen information in a memory. In one embodiment, a storage region may be set based on a touch point detected in the second region. Said at least a portion of the screen information may be included in the storage region. The scope of applicability of the present application will become more apparent from the detailed description below. Nevertheless, it should be understood that the detailed description and specific examples, although indicating preferred embodiments of the disclosure, are presented for illustrative purposes only, since various changes and modifications in the scope of the disclosure will become apparent to those skilled in the art from the detailed description. The accompanying drawings, which are included to provide a better understanding of the disclosure and which are incorporated herein, illustrate exemplary embodiments and, together with the description, serve to explain the principles of the disclosure.
[0012] The present invention will be described below in detail with reference to the accompanying drawings in which: Figure 1A is a block diagram of a mobile terminal according to the present invention; Figs. 1B and 1C are conceptual views of a mobile terminal according to an embodiment of the present invention; Fig. 2A is a perspective view of the front of a mobile terminal according to another embodiment of the present invention, and Fig. 2B is a perspective view of the rear of the mobile terminal of Fig. 2A; Fig. 3 is a flowchart illustrating a method of controlling a mobile terminal according to the present invention; Fig. 4 is a conceptual view for explaining the control method of Fig. 3; FIGS. 5A and 5B, 6A and 6B, 7A-7D, 8A-8C, 9A-9F and 10A-10C are conceptual views for explaining a method of providing an acquisition function, using a touch applied to a side display unit; Figs. 11, 12A-12C and 13 are conceptual views for explaining a method of providing various functions instead of an acquisition function, using a touch applied to a side display unit; Figs. 14A to 14D, 15A and 15B and 16A to 16C are conceptual views for explaining another embodiment of a method of providing an acquisition function, using a touch applied to a side display unit ; Figs. 17 (a) and 17 (b) are conceptual views for explaining another embodiment of a method of providing an acquisition function, using a touch applied to a side display unit; Figs. 18 (a) and 18 (b) are conceptual views for explaining a function of acquiring a plurality of screen information; Figs. 19A and 19B, 20A and 20B, 21A and 21B, 22 and 23A and 23B are conceptual views for explaining a method of controlling a camera function, using a touch applied to a side display unit; Figs. 24A-24C are conceptual views for explaining a method of controlling a function related to a moving image, using a touch applied to a side display unit; Figs. 25 (a) and 25 (b) are conceptual views for explaining a method of controlling an image output function and an image editing function, using a touch applied to a unit of image lateral display; and Fig. 26 is a conceptual view for explaining a method of controlling a search function, using a touch applied to a side display unit. The present invention will be described in detail hereinafter according to exemplary embodiments disclosed herein, with reference to the accompanying drawings. For the sake of brevity of the description with reference to the accompanying drawings, the same or equivalent components may have the same or similar reference numerals and their description is not repeated. In general, a suffix such as "module" or "unit" can be used to refer to elements or components. The use of such a suffix herein is simply to facilitate the description of the memoir, and the suffix itself is not intended to provide a special meaning or function. In the present disclosure, what is well known to those skilled in the art is generally omitted for the sake of brevity. The accompanying drawings are used to facilitate the understanding of various technical features and it should be appreciated that the embodiments described herein are not limited by the accompanying drawings. As such, the present disclosure should be understood to encompass all alterations, equivalencies and substitutions in addition to those particularly exemplified in the accompanying drawings. The mobile terminals described herein can be implemented using a variety of different types of terminals. Examples of such terminals include cell phones, smartphones, user equipment, laptops, digital broadcast terminals, personal digital assistants (PDAs), portable media players (PMPs), browsers, laptops ( PCs), slab PCs, tablet PCs, ultra books, portable devices (eg, smart watches, smart glasses, head-mounted displays (HMDs)) and the like. As a nonlimiting example only, particular types of mobile terminals will be described in detail below. Nevertheless, such teachings also apply to other types of terminals, such as the aforementioned types. In addition, these teachings can also be applied to immobile terminals, such as digital TVs, desktops and the like. Reference will be made hereinafter to FIGS. 1A to 1C, where FIG. 1A is a block diagram of a mobile terminal according to the present disclosure, and FIGS. 1B and 1C are conceptual views of an example of the mobile terminal. , in view in different senses. The mobile terminal 100 shown comprises components such as a wireless communication unit 110, an input unit 120, a detection unit 140, an output unit 150, an interface unit 160, a memory 170, a storage unit 180 control, and a power supply unit 190. It should be understood that the implementation of all the components shown is not a requirement and that more or fewer components can be implemented alternatively. With reference to FIG. 1A, the wireless communication unit 110 generally comprises one or more modules that enable communications such as wireless communications between the mobile terminal 100 and a wireless communication system, communications between the mobile terminal. 100 and another mobile terminal, or communications between the mobile terminal 100 and an external server. In addition, the wireless communication unit 110 generally includes one or more modules that connect the mobile terminal 100 to one or more networks. To facilitate such communications, the wireless communication unit 110 comprises one or more of a broadcast receiving module 111, a mobile communication module 112, a wireless Internet module 113, a module Short-range communication device 114 and location information module 115. Input unit 120 includes camera 121 for obtaining images or video, microphone 122, which is a type of communication device. audio input to input an audio signal, and a user input unit 123 (for example, a touch key, a push key, a mechanical key, a function key, and the like) to enable a user to enter information. Data (e.g., audio, video, image and the like) is obtained by the input unit 120 and can be analyzed and processed by the controller 180 according to device parameters, user commands and combinations thereof. The detection unit 140 is generally implemented using one or more sensors configured to detect internal information of the mobile terminal, the mobile terminal environment, user information and the like. For example, in FIG. 1A, it is shown that the detection unit 140 comprises a proximity sensor 141 and a lighting sensor 142. If this is desired, the detection unit 140 may alternatively or in addition , include other types of sensors or devices, such as a touch sensor, an acceleration sensor, a magnetic sensor, a gravity sensor, a gyroscope sensor, a motion sensor, an RGB sensor, an infrared sensor ( IR), a fingerprint sensor, an ultrasound sensor, an optical sensor (e.g., a camera 121), a microphone 122, a battery meter, an environmental sensor (for example, a barometer, a hygrometer, a thermometer, a radiation detection sensor, a thermal sensor and a gas sensor, among others), and a chemical sensor (e.g., an electronic nose, a health sensor, a biometric sensor, and the like) ), to name a few ns. The mobile terminal 100 may be configured to use information obtained from the detection unit 140, and in particular, information obtained from one or more sensors of the detection unit 140, and combinations of those -this. The output unit 150 is generally configured to output various types of information, such as audio, video, and touch outputs and the like. It is shown that the output unit 150 comprises a display unit 151, an audio output module 152, a haptic module 153, and an optical output module 154. The display unit 151 may comprise a laminated structure or an integrated structure with a touch sensor to facilitate a touch screen. The touch screen can provide an output interface between the mobile terminal 100 and a user, and it can also function as a user input unit 123 which provides an input interface between the mobile terminal 100 and the user. 'user. The interface unit 160 serves as an interface with various types of external devices that can be coupled to the mobile terminal 100. The interface unit 160 may for example comprise wired or wireless ports, power supply ports. external, wired or wireless data ports, memory card ports, ports for connecting a device with an identification module, input / output (I / O) ports, video I / O ports , earphone ports and similar items. In some cases, the mobile terminal 100 may perform matching control functions associated with an attached external device, in response to the link between the external device and the interface unit 160. The memory 170 is generally implemented to memorize data to support various functions or features of the mobile terminal 100. For example, the memory 170 may be configured to store application programs executed in the mobile terminal 100, data or instructions for operations of the mobile terminal 100 , and similar elements. Some of these application programs can be downloaded from an external server via wireless communication. Other application programs may be installed in the mobile terminal 100 at the time of manufacture or shipment, which is generally the case for basic functions of the mobile terminal 100 (for example, receiving a call, passing a call, receive a message, send a message, and similar items). It is common for the application programs to be stored in the memory 170, installed in the mobile terminal 100, and executed by the controller 180 to perform an operation (or function) for the mobile terminal 100. The controller 180 generally operates to control the overall operation of the mobile terminal 100, in addition to the operations associated with the application programs. The controller 180 may provide or process information or functions appropriate for a user by processing signals, data, information and the like which are input or output from the various components shown in FIG. 1A, or by activating application programs stored in the memory 170. In one example, the controller 180 controls some or all of the components illustrated in FIGS. 1A to 1C depending on the execution of an application program which has been stored in the memory 170. The power supply unit 190 may be configured to receive external energy or provide internal power to supply the appropriate power necessary for the operation of the elements and components included in the terminal. The power supply unit 190 may comprise a battery, and the battery may be configured to be integrated into the body of the battery. terminal, or configured to be detachable from the terminal body.
[0013] With reference to FIG. 1A, various components shown in this figure will be described in detail below. With respect to the wireless communication unit 110, the broadcast receiving module 111 is generally configured to receive a broadcast signal and / or information associated with a broadcast from an external broadcast management entity. via a broadcast channel. The broadcast channel may include a satellite channel, a terrestrial channel, or both. In some embodiments, two or more broadcast receiving modules 111 may be used to facilitate the simultaneous reception of two or more broadcast channels, or to support switching between broadcast channels.
[0014] The mobile communication module 112 may transmit and / or receive wireless signals to and / or from one or more network entities. Typical examples of a network entity are a base station, an external mobile terminal, a server, and the like. Such network entities are part of a mobile communication network that is built according to technical standards or communication methods for mobile communications (eg, global mobile communication system (GSM), code division multiple access ( CDMA), Code Division Multiple Access 2000 (CDMA2000), Broadband CDMA (WCDMA), High Speed Downlink Packet Access (HSDPA), Long Term Evolution (LTE), and similar elements). Examples of wireless signals transmitted and / or received via the mobile communication module 112 include audio calling signals, video calling (telephony) signals, or various data formats for supporting communications. texts and multimedia messages.
[0015] The wireless Internet module 113 is configured to facilitate access to the wireless Internet. This module can be coupled, internally or externally, to the mobile terminal 100. The wireless Internet module 113 can transmit and / or receive wireless signals via communication networks according to wireless Internet technologies. Examples of such wireless Internet access are in particular wireless local area network (WLAN), wireless fidelity (Wi-Fi), Wi-Fi Direct, digital living network alliance (DLNA), wireless broadband (WiBro), global interoperability for microwave access (WiMAX), high speed downlink packet access (HSDPA), long term evolution (LTE) and similar elements. The wireless Internet module 113 may transmit / receive data based on one or more of these wireless Internet technologies as well as other Internet technologies.
[0016] In some embodiments, when wireless Internet access is implemented based on, for example, WiBro, HSDPA, GSM, CDMA, WCDMA, LTE and the like, as part of a mobile communication network. , the wireless Internet module 113 performs such wireless Internet access. As such, the Internet module 113 can cooperate with or operate as the mobile communication module 112. The short-range communication module 114 is configured to facilitate short-range communications. Suitable technologies for implementing such short-range communications are in particular BLUETOOTHTM, radio frequency identification (RFID), infrared data association (IrDA), ultra-wideband (UWB), ZigBee, near-field communication (NFC), fidelity Wireless (Wi-Fi), Wi-Fi Direct, Universal Wireless Serial Bus (USB Wireless), and similar items. In general, the short-range communication module 114 supports wireless communications between the mobile terminal 100 and a wireless communication system, communications between the mobile terminal 100 and another mobile terminal 100, or communications between the mobile terminal 100 and a mobile terminal 100, or communications between the mobile terminal 100 and another mobile terminal 100, or communications between the mobile terminal 100. mobile terminal and a network in which there is another mobile terminal 100 (or an external server), via wireless networks. An example of a wireless network is a wireless personal network. In some embodiments, another mobile terminal (which may be configured similarly to the mobile terminal 100) may be a portable device, for example a smart watch, smart glasses, or a head-mounted display (HMD), which is capable of exchanging data with the mobile terminal 100 (or cooperating otherwise with the mobile terminal 100). The short-range communication module 114 can detect or recognize the portable device, and allow communication between the portable device and the mobile terminal 100. In addition, when the detected portable device is a device that is authenticated to communicate with the mobile terminal 100, the controller 180, for example, can cause processed data to be transmitted in the mobile terminal 100 to the portable device via the short-range communication module 114. A user of the portable device can then use the data. processed in the mobile terminal 100 on the portable device. For example, when a call is received in the mobile terminal 100, the user can answer the call using the portable device. Similarly, when a message is received in the mobile terminal 100, the user can view the received message using the portable device. The location information module 115 is generally configured to detect, calculate, derive or otherwise identify a position of the mobile terminal. In one example, the location information module 115 includes a global positioning system (GPS) module, a Wi-Fi module, or both. If desired, the location information module 115 may alternatively or additionally operate with one or more other modules of the wireless communication unit 110 to obtain data relating to the position of the mobile terminal. In one example, when the mobile terminal uses a GPS module, a position of the mobile terminal can be acquired using a signal sent from a GPS satellite. In another example, when the mobile terminal uses the Wi-Fi module, a position of the mobile terminal can be acquired based on information about a wireless access point (AP) that transmits or receives a wireless signal. to or from the Wi-Fi module. The input unit 120 may be configured to allow various types of input to the mobile terminal 120. Examples of such an input include an audio input, an input of image, a video input, a data input and a user input. An image input and a video input are often obtained using one or more cameras 121. Such cameras 121 can process frames of images of photos or videos obtained by image sensors in an acquisition mode. video or image. The processed image frames may be displayed on the display unit 151 or stored in the memory 170. In some cases, the cameras 121 may be arranged in a matrix configuration to allow a plurality of images having various angles or various focal points to be input to the mobile terminal 100. In another example, the cameras 121 may be located in a stereoscopic arrangement for acquiring left and right images to implement a stereoscopic image.
[0017] The microphone 122 is generally implemented to allow audio input into the mobile terminal 100. The audio input may be processed in a variety of ways according to a function performed in the mobile terminal 100. If desired, the microphone 122 may include matching noise suppression algorithms to suppress unwanted noise generated during the reception of external audio. The user input unit 123 is a component that allows a user to make an entry. Such user input may allow the controller 180 to control the operation of the mobile terminal 100. The user input unit 123 may include one or more of a mechanical input element (e.g. , a key, a button located on a front surface and / or a back surface or a side surface of the mobile terminal 100, a dome-shaped switch, a scroll wheel, a wheel and the like), or a touch input , among others. In one example, the touch input can be a virtual key or a function key, which is displayed on a touch screen via software processing, or a touch key that is located on the mobile terminal at a location other than the touch screen. Furthermore, the virtual key or the visual key can be displayed on the touch screen in various forms, for example graphic, text, icon, video or a combination thereof. The detection unit 140 is generally configured to detect one or more internal information of the mobile terminal, mobile terminal environment information, user information or similar information. The controller 180 generally cooperates with the sending unit 140 to control the operation of the mobile terminal 100 or to execute a data processing, a function or an operation associated with an application program installed in the mobile terminal on the Based on the detection provided by the detection unit 140. The detection unit 140 may be implemented using any of a variety of sensors, some of which will be described in detail hereinafter.
[0018] The proximity sensor 141 may comprise a sensor for detecting the presence or absence of an object approaching a surface, or an object located near a surface, using a magnetic field, infrared rays or similar elements without mechanical contact. The proximity sensor 141 may be arranged at an inner region of the mobile terminal covered by the touch screen, or near the touch screen. The proximity sensor 141 may for example comprise any one of a transmissive type photoelectric sensor, a direct reflection type photoelectric sensor, a mirror reflection type photoelectric sensor, a high oscillation proximity sensor. frequency, a capacitance type proximity sensor, a magnetic type proximity sensor, an infrared proximity sensor, and the like. When the touch screen is implemented as a capacitance type, the proximity sensor 141 can detect the proximity of a pointer to the touch screen by changes in an electromagnetic field in response to an approach to the touch screen. an object with a conductivity. In this case, the touch screen (touch sensor) can also be categorized as a proximity sensor. In the present, reference is often made to the term "proximity touch" to indicate a scenario in which a pointer is positioned to be close to the touch screen without touching it. In the present, reference is often made to the term "contact touch" to indicate a scenario in which a pointer is in physical contact with the touch screen. For the position corresponding to the proximity touch of the pointer relative to the touch screen, such a position corresponds to a position at which the pointer is perpendicular to the touch screen. The proximity sensor 141 can detect a proximity touch, and proximity touch patterns (e.g., distance, direction, speed, time, position, travel status, and the like). In general, the controller 180 processes data corresponding to proximity taps and proximity tap patterns detected by the proximity sensor 141, and causes visual information to be output to the touch screen. In addition, the controller 180 may control the mobile terminal 100 to perform different operations or to process different data depending on whether a touch at a point on the touch screen is a proximity touch or a touch. A touch sensor can detect a touch applied to the touch screen, such as a display unit 151, using any of a variety of touch methods. Examples of such tactile methods are a resistive type, a capacitive type, an infrared type and a magnetic field type, among others. In one example, the touch sensor may be configured to convert pressure changes applied to a specific portion of the display unit 151, or convert a capacitance occurring at a specific portion of the display unit 151, 10 to electrical input signals. The touch sensor can also be configured to detect not only an affected position and an affected area, but also touch pressure and / or touch ability. A touch object is generally used to apply a touch input to the touch sensor. Examples of typical touch objects are a finger, a touch pen, a stylus, a pointer, or the like. When a touch input is detected by a touch sensor, corresponding signals can be transmitted to a touch controller. The touch controller can process the received signals and then transmit corresponding data to the controller 180. The controller 180 can therefore detect the region of the display unit 151 that has been touched. Here, the touch controller may be a separate component of the controller 180, the controller 180 and combinations thereof. In some embodiments, the controller 180 may perform the same or different commands depending on a type of touch object that touches the touch screen or a touch key provided in addition to the touch screen. It can be decided whether to execute the same command or a different command depending on the object that provides a touch input based on for example a current operating state of the mobile terminal 100 or a program of operation. application running. The touch sensor and the proximity sensor may be used individually or in combination to detect various types of touch. Such touches include a short touch (or a tap), a long touch, a multi-touch, a slip feel, a touch of movement, a nip pinch feel, a widening pinch feel, a swipe feel , a touch of passage and similar elements. If desired, an ultrasound sensor may be implemented to recognize positional information relating to a touch object using ultrasonic waves. The controller 180 may for example calculate a position of a wave generation source based on information detected by an illumination sensor and a plurality of ultrasonic sensors. Since light is much faster than ultrasonic waves, the time it takes for the light to reach the optical sensor is much shorter than the time it takes the ultrasonic waves to reach the ultrasonic sensor. The position of the wave generation source can be calculated using this fact. For example, the position of the wave generation source can be calculated using the time difference from the time it takes the ultrasonic wave to reach the sensor on the basis of the light as a signal of reference.
[0019] The camera 121 generally comprises at least one camera sensor (CCD, CMOS, etc.), a photosensor (or image sensors) and a laser sensor. The implementation of the camera 121 with a laser sensor may allow the detection of a touch of a physical object with respect to a 3D stereoscopic image. The photosensor may be laminated or superimposed on the display. The photosensor can be configured to scan the movement of the physical object near the touch screen. In detail, the photosensor may include photodiodes and transistors in rows and columns to scan the content received at the photosensor using an electrical signal that changes depending on the amount of light applied. The photosensor can calculate the coordinates of the physical object according to the variation of the light to obtain position information of the physical object. The display unit 51 is generally configured to output processed information to the mobile terminal 100. For example, the display unit 151 may display run screen information of an application program. executing at the mobile terminal 100 or the user interface (UI) and graphical user interface (GUI) information in response to the execution screen information.
[0020] In some embodiments, the display unit 151 may be implemented as a stereoscopic display unit for displaying stereoscopic images. A typical stereoscopic display unit may employ a stereoscopic display scheme such as a stereoscopic schema (eyeglass scheme), an autostereoscopic scheme (schematic without glasses), a projection scheme (holographic scheme) or a similar scheme. The audio output module 152 is generally configured to output audio data. Such audio data can be obtained from any one of a number of different sources, so that the audio data can be received from the wireless communication unit 110 or can be stored in the memory. Memory 170. The audio data may be output in modes such as a signal receiving mode, a calling mode, a recording mode, a voice recognition mode, a broadcast receiving mode, or a similar mode. The audio output module 152 may provide an audible output relating to a particular function (eg, a call signal receiving sound, a message receiving sound, etc.) performed by the mobile terminal 100. audio output 152 may also be implemented in the form of a receiver, a speaker, a buzzer or the like. A haptic module 153 may be configured to generate various tactile effects that a user feels, perceives or otherwise receives. A typical example of a tactile effect generated by the haptic module 153 is a vibration. The force, pattern and other similar characteristics of the vibration generated by the haptic module 153 may be controlled by a user selection or may be adjusted by the controller. For example, the haptic module 153 can deliver different vibrations in a combined manner or in a sequential manner. In addition to vibration, the haptic module 153 may generate various other tactile effects, including a pacing effect such as a vertically moving pin arrangement for contacting the skin, a spraying force, or a suction force. air through a jet orifice or a suction opening, a touch of the skin, an electrode contact, an electrostatic force, an effect by reproducing the sense of cold and heat by using an element capable of absorbing or generate heat, and similar elements.
[0021] The haptic module 153 can also be implemented to allow the user to feel a tactile effect through a muscular sensation, such as the fingers or the arm of the user, and transferring the tactile effect by direct contact. Two or more haptic modules 153 may be provided according to the particular configuration of the mobile terminal 100. An optical output module 154 may output a signal to indicate event generation using light from a light source. Examples of events generated in the mobile terminal 100 may include message reception, call waiting reception, missed call, alarm, scheduling notification, e-mail reception information via an application, and similar elements. A signal delivered by the optical output module 154 may be implemented such that the mobile terminal transmits monochromatic light or light with a plurality of colors. The signal output may be terminated when the mobile terminal detects that a user has viewed the generated event, for example. The interface unit 160 serves as an interface for external devices to be connected to the mobile terminal 100. For example, the interface unit 160 can receive data transmitted from an external device, receive an energy to be transmitted to 20 elements and components within the mobile terminal 100, or transmit internal data from the mobile terminal 100 to such an external device. The interface unit 160 may include wired or wireless headphone ports, external power supply ports, wired or wireless data ports, memory card ports, ports for connecting a device having a identification module, audio input / output (I / O) ports, video I / O ports, headphone ports, or the like. The identification module may be a chip that stores various information to authenticate the authority to use the mobile terminal 100 and may include a user identity module (UIM), a subscriber identity module 30 ( SI1V1), a universal subscriber identity module (USIM), and the like. In addition, the device comprising the identification module (which is referred to hereinafter as "identification device") may take the form of a smart card. Therefore, the identification device can be connected to the terminal 100 via the interface unit 160. When the mobile terminal 100 is connected to an external docking station, the interface unit 160 can serve as a passageway for providing the power supply from the docking station to the mobile terminal 100 or it can be used as a gateway to allow the transfer of various control signals entered by the user to the mobile terminal. Various control signals or the power supply of the docking station may function as signals to recognize that the mobile terminal is properly mounted on the docking station.
[0022] The memory 170 may store programs to support operations of the controller 180 and store input / output data (eg a directory, messages, photos, videos, etc.). The memory 170 can store data relating to various vibration and audio patterns that are delivered in response to touch inputs on the touch screen.
[0023] The memory 170 may comprise one or more types of storage media, including a flash memory, a hard disk, a solid state disk, a semiconductor disk, a multimedia microcard, a card memory (for example, SD or DX memory, etc.), random access memory (RAM), static random access memory (SRAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, a magnetic disk, an optical disk and the like. The mobile terminal 100 may also operate in connection with a network storage device that performs the function of storing memory 170 on a network, such as the Internet.
[0024] The controller 180 may generally control the general operations of the mobile terminal 100. For example, the controller 180 may activate or deactivate a lockout state to prohibit a user from entering a control command with respect to applications when a mobile terminal status meets a preset condition.
[0025] The controller 180 may also perform the control and processing associated with voice calls, data communications, video calls, and the like, or perform pattern recognition processing to recognize a handwritten input or a speech input. drawing done on the touch screen respectively in the form of characters or images. In addition, the controller 180 may control a component or combination of these components to implement various exemplary embodiments disclosed herein.
[0026] The power supply unit 190 receives external energy or provides internal energy and supplies the appropriate energy necessary for the operation of respective elements and components included in the mobile terminal 100. The power supply unit 190 may comprise a battery, which is generally rechargeable or releasably coupled to the terminal body for charging.
[0027] The power supply unit 190 may include a link port. The link port may be configured as an example of the interface unit 160 to which an external charger is electrically connected to provide the energy required to recharge the battery. In another example, the power supply unit 190 may be configured to recharge the battery in a wireless manner without using the link port. In this example, the power supply unit 190 can receive energy transferred by an external wireless energy transmitter, using at least one of an inductive coupling method which is based on induction magnetic or a magnetic resonance coupling method that is based on magnetic resonance. Various embodiments described herein may be implemented in a computer readable medium, a machine readable medium or a similar medium using for example software, hardware or a combination thereof.
[0028] With reference to FIGS. 1B and 1C, the mobile terminal 100 will be described below with respect to a bar-type terminal body. Nevertheless, the mobile terminal 100 may alternatively be implemented in various configurations. Examples of such configurations are in particular a watch, a clip, glasses, a folding type, a folding type, a sliding type, a tilt type, and a pivot type in which two or more bodies are combined so as to ability to move relative to each other, and combinations thereof. The following description often focuses on a particular type of mobile terminal (for example, a bar type, a watch, glasses, and similar types).
[0029] Nevertheless, such teachings concerning a particular type of mobile terminal also apply in general to other types of mobile terminals. The mobile terminal 100 generally comprises a housing (for example, a chassis, a housing, a cover, and the like) forming the appearance of the terminal. In the present embodiment, the housing is constituted using a front housing 101 and a rear housing 102. Various electronic components are incorporated in a space formed between the front housing 101 and the rear housing 102. At least one central housing can in addition to being positioned between the front housing 101 and the rear housing 102.
[0030] The display unit 151 is shown located on the front side of the terminal body for providing information. As illustrated, a window 151a of the display unit 151 may be mounted on the front housing 101 to form the front surface of the terminal body with the front housing 101. In some embodiments, electronic components may also be provided. The examples of such electronic components include a detachable battery 191, an identification module, a memory card and the like. A rear cover 103 is shown covering the electronic components, and this cover can be releasably coupled to the rear housing 102. Therefore, when the rear cover 103 is detached from the rear housing 102, the electronic components mounted in the rear housing 102 are exposed to the outside. As illustrated, when the rear cover 103 is coupled to the rear housing 102, a side surface of the rear housing 102 is partially exposed. In some cases, during the coupling, the rear housing 102 may also be completely protected by the rear cover 103. In some embodiments, the rear cover 103 may include an opening for exposing a camera 121b or a video module to the outside. 152b audio output. The housings 101, 102, 103 may be formed by injection molding of synthetic resin or they may be made of a metal, for example stainless steel (STS), aluminum (Al), titanium (Ti ) or similar elements. In an alternative to the example in which the plurality of housings constitute an interior space for receiving components, the mobile terminal 100 may be configured so that a housing forms the interior space. In this example, a mobile terminal 100 having a single body is formed such that a synthetic resin or metal extends from a side surface to a back surface. If desired, the mobile terminal 100 may include a watertight unit (not shown) to prevent water from entering the terminal body. For example, the watertight unit may include a watertight member that is located between the window 151a and the front housing 101, between the front housing 101 and the rear housing 102, or between the rear housing 102. and the back cover 103, for hermetically sealing an interior space when these housings are coupled. The display unit 151, the first audio output module 152a, the second audio output module 152b, the proximity sensor 141, the illumination sensor 142, the optical output unit 154, the first camera 121a, the second camera 121b, the first handling unit 123a, the second handling unit 123b, the microphone 122, the interface unit 160, etc. can be provided in the mobile terminal 100. As shown in FIGS. 1B and 1C, it is assumed that the display unit 151, the first audio output module 152a, the proximity sensor 141, the sensor illumination 142, the optical output unit 154, the first camera 121a and the first handling unit 123a are disposed on a front surface of the body. It is assumed that the second manipulation unit 123b, the microphone 122 and the interface unit 160 are disposed on side surfaces of the body. It is also assumed that the second audio output module 152b and the second camera 12b are disposed on a back surface of the body. Figures 1B and 1C show certain components arranged on the mobile terminal. Nevertheless, it should be understood that alternative arrangements are possible within the scope of the teachings of the present disclosure. Some components may be omitted or arranged differently. For example, the first handling unit 123a may be located on another surface of the terminal body, and the second audio output module 152b may be located on the side surface of the terminal body. The display unit 151 delivers processed information in the mobile terminal 100. For example, the display unit 151 may display information on an execution screen of a controlled application program in the mobile terminal 100 , user interface (UI) information, or graphical user interface (GUI) information. The display unit 151 may be implemented using one or more suitable display devices. Examples of such suitable display devices include a liquid crystal display (LCD), a thin film transistor (TFT-LCD) liquid crystal display, an organic light-emitting diode (OLED), a flexible display, a three-dimensional display (3D), an electronic ink display and combinations thereof. The display unit 151 may be implemented using two display devices, which may use the same display technology or different display technologies. For example, a plurality of display units 151 may be arranged on one side being spaced apart from one another or integrated with each other, or these devices may be arranged on different surfaces. . The display unit 151 may also include a touch sensor that detects a touch input received at the display unit. When a touch is input to the display unit 151, the touch sensor can be configured to detect that touch and the controller 180, for example, can generate a control command or other signal corresponding to the touch. The content that is entered by touch can be a text or numeric value, or a menu item that can be specified or designated in various modes. The touch sensor may be configured as a film having a tactile pattern, disposed between the window 151a and a display on a rear surface of the window 151a, or a wire which is structured directly on the rear surface of the window 151a. Alternatively, the touch sensor may be an integral part of the display. For example, the touch sensor may be disposed on a substrate of the display or within the display.
[0031] The display unit 151 may also constitute a touch screen with the touch sensor. Here, the touch screen can serve as a user input unit 123 (see Figure 1A). Therefore, the touch screen can replace at least some of the functions of the first handling unit 123a.
[0032] The first audio output module 152a may be implemented as a receiver for transmitting a call sound to a user's ears, and the second audio output module 152b may be implemented in the form of a receiver. speaker to deliver voice audio, alarm sounds, multimedia audio reproduction and the like. The window 151a of the display unit 151 generally includes an aperture to allow passage of the audio generated by the first audio output module 152a. An alternative is to pass the audio along an assembly gap between the structural bodies (for example, a spacing between the window 151a and the front housing 101). In this case, an independently formed hole for delivering audio sounds is not necessarily visible or otherwise hidden in appearance, further simplifying the appearance and fabrication of the mobile terminal 100. The optical output module 154 can be configured to output light to indicate event generation. Examples of such events include message reception, call waiting reception, missed call, alarm, scheduling notification, e-mail reception, receipt of information via An Application, and Similar Elements When a user has viewed a generated event, the controller may command the optical output unit 154 to turn off the light output. The first camera 121a can process frames of images as still or moving images obtained by the image sensor in an acquisition mode or in a video call mode. The processed image frames can then be displayed on the display unit 151 or stored in the memory 170. The first and second handling units 123a and 123b are examples of the user input unit 123. , which may be manipulated by a user to provide an input to the mobile terminal 100. Reference may also be made to the first and second handling units 123a and 123b as a handling portion. The first and second handling units 123a and 123b may employ any tactile method that allows the user to perform manipulation such as touching, pushing, scrolling or similar manipulation. The first and second handling units 123a and 123b may also employ any non-tactile method that allows the user to perform manipulation such as proximity touching, moving, or similar manipulation. Figure 1B illustrates the first manipulation unit 123a as a touch key, but possible alternatives include a mechanical key, a push key, a touch key, and combinations thereof. Inputs received at the first and second handling units 123a and 123b can be used in a variety of ways. For example, the first handling unit 123a may be used by the user to provide menu entry, a return key at the beginning, a cancel, a search, or the like, and the second handling unit 123b may be used by the user to provide an input for controlling a volume level output from the first or second audio output module 152a or 152b, to switch to a touch recognition mode of the display unit 151 or 15 similar elements. In another example of the user input unit 123, a rear input unit (not shown) may be located on the rear surface of the terminal body. The rear input unit may be manipulated by a user to provide an input to the mobile terminal 100. The input may be used in a variety of different ways. For example, the rear input unit may be used by the user to provide an input for power on / off, start, end, scroll, volume level adjustment delivered by the first or the second audio output module 152a or 152b, the switch to a touch recognition mode of the display unit 151, and the like. The rear input unit 25 may be configured to allow touch input, push input, or combinations thereof. The rear input unit may be located to overlap the display unit 151 on the front side in the thickness direction of the terminal body. In one example, the rear input unit may be located on an upper end portion of the rear side of the terminal body so that a user can easily manipulate it with the index when the user takes the body. terminal of a hand. Alternatively, the rear input unit may be positioned at virtually any location on the rear side of the terminal body.
[0033] Embodiments comprising the rear input unit may implement some or all of the features of the first handling unit 123a in the rear input unit. In situations in which the first handling unit 123a is omitted from the front side, the display unit 151 may have a larger screen. In another alternative, the mobile terminal 100 may include a fingerprint sensor that scans a user's fingerprint. The controller 180 may then use fingerprint information detected by the fingerprint sensor as part of an authentication procedure. The fingerprint sensor may also be installed in the display unit 151 or may be implemented in the user input unit 123. The microphone 122 is shown at one end of the mobile terminal 100, but other locations are possible. If desired, a plurality of microphones may be implemented in an arrangement for receiving stereo sounds. The interface unit 160 may serve as a path for the mobile terminal 100 to interface with external devices. For example, the interface unit 160 may include one or more of a link terminal for connection to another device (e.g., headphones, speaker, or the like), a port for an near-field communication (for example, an Infrared Data Association (IrDA) port, a Bluetooth port, a wireless LAN port, and the like), or a power supply terminal for providing power to the mobile terminal 100. The interface unit 160 may be implemented in the form of a housing for receiving an external card, such as a subscriber identification module (SEM), an identity module of subscriber (UIM) or a memory card for storing information. The second camera 121b is shown located on the rear side of the terminal body and includes an image acquisition direction which is substantially opposite to the image acquisition direction of the first camera unit 121a. If desired, the second camera 121a may be alternatively located at other locations or may be made movable to have a different image acquisition direction than the one shown.
[0034] The second camera 121b may include a plurality of lenses arranged along at least one line. The plurality of lenses may also be arranged in a matrix configuration. Cameras can be referred to as "network cameras". When the second camera 121b is implemented as a network camera, images can be acquired in various ways using the plurality of lenses and the images are of better quality. As shown in FIG. 1C, a flash 124 is adjacent to the second camera 121b. When an image of a subject is acquired with the camera 121b, the flash 124 may illuminate the subject. As shown in FIG. 1B, the second audio output module 152b may be located on the terminal body. The second audio output module 152b may implement stereophonic sound functions in connection with the first audio output module 152a, and it may also be used to implement a speaker mode for call communication. . At least one antenna for wireless communication may be located on the terminal body. The antenna can be installed in the terminal body or formed by the housing. For example, an antenna that configures a portion of the broadcast receiving module 111 may be retractable into the terminal body. Alternatively, an antenna may be formed using a film attached to an inner surface of the back cover 103, or a housing that includes a conductive material. A power supply unit 190 for supplying power to the mobile terminal 100 may include a battery 191, which is mounted in the terminal body or is detachably coupled outside the terminal body. . The battery 191 can receive power via a power source cable connected to the interface unit 160. Similarly, the battery 191 can be recharged in a wireless manner using a wireless charger. Wireless charging can be implemented by magnetic induction or electromagnetic resonance. Rear cover 103 is shown coupled to rear housing 102 to shield battery 191 to prevent separation of battery 191 and to protect battery 191 against external impact or foreign objects. When the battery 191 is detachable from the terminal body, the rear cover 103 may be releasably coupled to the rear housing 102. An accessory to protect an appearance or to assist or expand the functions of the mobile terminal 100 may also be provided on the mobile terminal 100. An example of an accessory is a case or a facade for covering or receiving at least one surface of the mobile terminal 100. The case or facade can cooperate with the display unit 151 to extend the function of the terminal mobile 100. Another example of an accessory is a touch pen to help or extend a touch input on a touch screen.
[0035] The mobile terminal according to the present invention may have various designs. A mobile terminal having a side display unit, as a structural change and structural modification, and a user interface using the side display unit will be explained hereinafter. Fig. 2A is a perspective view of the front of a mobile terminal according to another embodiment of the present invention, and Fig. 2B is a perspective view of the rear of the mobile terminal of Fig. 2A. The mobile terminal 200 according to the present disclosure is a bar-type mobile terminal. Nevertheless, the present disclosure is not limited to this, but it can be applied to a sliding type, in which two or more bodies are coupled to each other so as to effect relative movement, a collapsible type. , a tilting type, a pivot type or other similar type. A housing (housing, housing, cover, etc.) forming the appearance of the body 200 may include a front housing 201 and a rear housing 202 (see Figures 3A and 3B). A space formed by the front housing 201 and the rear housing 202 may receive various components therein. At least one intermediate housing may further be disposed between the front housing 201 and the rear housing 202. Such housings may be formed by synthetic resin injection molding, or they may be formed using a metallic material such as stainless steel (STS) or titanium (Ti). A display unit 251, an audio output unit 252, a camera 221, etc. may be disposed at the front housing 201. An interface unit 270, etc. may be disposed on side surfaces of the front housing 201 and the rear housing 202. The display unit 251 occupies most of a main surface of the front housing 201. This means that the display unit is disposed on a front surface of the mobile terminal, and is configured to display visual information. The display unit 251 is formed not only on a front surface of the mobile terminal, but also on other surfaces of the mobile terminal. More specifically, the display unit 251 is provided with a first region 252 disposed on the front surface, and a second region 253 extending from the first region 252 and disposed on a side surface of the body. The side surface may be a surface visible to a user when the user is looking at the mobile terminal by the side. At least a portion of the second region 253 may be disposed on the front surface. For example, the second region 253 may be formed on the side surface and on the front surface. It is determined whether the second region 253 is visible from the front surface according to a structure formed by the first region 252 and the second region 253 on the display unit 251. For example, a window disposed on a top surface of the display unit 251 is formed so that two side surfaces thereof are folded. In such a configuration, the appearance of the front surface and side surfaces is formed by the window. Thus, the first region 252 and the second region 253 can be connected to each other without any boundary between them. In this case, the display unit 251 may be formed in a folded manner, and may be provided with a display device mounted in correspondence with the window.
[0036] In another example, the display unit 251 may be a flexible display unit. The flexible display unit includes a display that can be folded, twisted, folded, and wound by an external force. The flexible display unit may comprise a general flexible display and electronic paper. General soft display means a light, non-fragile display formed on a thin and flexible substrate, which maintains features of the conventional flat display. The thin and flexible substrate can be folded, folded or rolled up like paper.
[0037] The electronic paper is implemented by applying general ink characteristics and it may be different from a general flat display in that the reflected light is used. Electronic paper can change images or text using rotating beads or electrophoresis using capsules.
[0038] With such a flexible material, the terminal body can be configured so that two side surfaces of a display unit are folded. The audio output module 252 and the camera 221 are disposed in a region adjacent to one of two ends of the display unit 251, and a front input unit (not shown) and a microphone 222 can be arranged in one region adjacent to another of the two ends of the display unit 251. The front input unit is an example of the user input unit 230 (see FIG. can include a plurality of handling units. Handling units may be referred to as handling portions, and may include any type of portions that can be touch manipulated by a user. The display unit 251 may constitute a touch screen with a touch sensor. In this case, the touch screen may be a user input unit. In such a configuration, the mobile terminal may be devoid of front input unit on the front surface thereof. In this case, the mobile terminal may be configured so that an input to the terminal body can be made through the display unit 251 and a rear input unit 232, as will be explained later. With reference to FIG. 2B, a camera 221 'may be further provided on the rear housing 202. The camera 221' faces in a direction opposite to a direction in which the camera 221 faces (see FIG. 2A), and it may have pixels different from the camera 221. For example, the camera 221 may operate with relatively lower pixels (lower resolution). Thus, the camera 221 may be useful when a user can take a picture of his face and send it to a correspondent during a video call. On the other hand, the camera 221 'can operate with relatively higher pixels (higher resolution) to allow a user to obtain higher quality images for later use. The cameras 221 and 221 'can be installed at the body so as to be able to turn or to be able to exit. In addition, a flash and a mirror may be disposed adjacent to the camera 221 '. The flash operates in connection with the camera 221 'when a picture is taken with the camera 221'. The mirror may cooperate with the camera 221 'to allow a user to photograph themselves in self-portrait mode. An audio output unit (not shown) may further be provided at the rear housing 202. The audio output unit may implement a stereo function with the audio output unit 252 (see Fig. 2A) and can be used to make a call in a speaker mode. The audio output unit 252 (the first audio output unit) may be formed on the front surface of the mobile terminal as a receiver, and a second audio output unit may be formed on the rear surface of the mobile terminal as loud speaker. Nevertheless, the present invention is not limited thereto. Indeed, the second audio output unit may be provided on a side surface of the mobile terminal. A power supply unit 290 for supplying power to the mobile terminal 200 is mounted on the body. The power supply unit 290 may be mounted in the body or it may be detachably mounted to the body. A rear entry unit 232 may be disposed on a rear surface of the body. The rear input unit 232 may be positioned below the camera 221 '. Commands input through the rear input unit 232 can be set in a variety of ways. For example, the rear input unit 232 is configured to input commands such as power on / off, start, end, scroll or similar commands. The rear input unit 232 is also configured to input commands for controlling a sound level output from the audio output unit 252, or commands to convert the current mode of the display unit 251 to a mode tactile recognition. Nevertheless, the present invention is not limited thereto. Indeed, the mobile terminal may be provided with one of the front input unit and the rear input unit 232, or both.
[0039] As mentioned above, the controller 180 (see Fig. 1A) can control functions of the mobile terminal, using a display disposed on a side surface of the mobile terminal (a side display unit). A method of controlling the function of the mobile terminal using a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. An acquisition function, among various functions that can be provided using a side display unit, will be explained in detail hereinafter. A first region means a display region disposed on a front surface of the mobile terminal (i.e., the front display unit), the display region having a front surface, side surfaces, and a surface back. The second region means a region disposed on a side surface of the mobile terminal (i.e., the side display unit). The display unit 251 of the mobile terminal may include a first region 261 formed on the front surface of the display unit 251, a second region 262 formed on a right side based on the first region 261, and a third region 263 formed on a left side based on the first region 261. The display unit 251 of the present invention is provided with two display regions on the right side and on the left side. Nevertheless, the display unit 251 may further include a side display unit on one of the right side and the left side based on the first region 261 formed on the front surface of the display unit 251. For convenience, it is assumed that the display unit 251 of the present invention includes display units (display regions) on the front surface, the left side surface and the right side surface. It is thus clear to those skilled in the art that the present invention can be applied to a mobile terminal comprising a display region (a display unit) only on one of the left side surface and the surface. right side. In the present invention, a front display unit (first region) bears the reference numeral "261" and a side display unit (second region or third region) bears the reference number "262" or "263". . In the case where the front surface and the side surfaces can also be used, the display unit 251 will be explained hereinafter instead of the front surface and the side surface.
[0040] Reference can be made to an acquisition function performed by the mobile terminal according to the present invention as a "screen capture function". The acquisition function means that screen information supplied to the display unit 251 is stored in the memory in an image file format. The image includes a still image and a moving image. The acquisition function may also mean that screen information supplied to the display unit 251 is stored in the memory in a file format other than an image file format. A method for providing an acquisition function in the mobile terminal 10 according to the present invention will be explained in detail hereinafter with reference to the accompanying drawings. Fig. 3 is a flowchart illustrating a method for controlling a mobile terminal according to the present invention, and Fig. 4 is a conceptual view for explaining the control method of Fig. 3. Referring to Figs. Screen 300 is supplied to the first region (front display unit 261) disposed on the front surface of the mobile terminal (S310). The screen information may be information of various types that may be output to the display unit 251. For example, the screen information may be an application run screen, a menu screen, a lock screen or a home screen. A type of screen information delivered to the display unit 251 may be determined by a selection of a user or under the control of the controller 180. In a state in which the screen information 300 have been delivered in the first region (front display unit 261) (S310), if a preset type of touch is applied in the second region (side display unit 262), the controller 180 acquires at the first region (front display unit 261) (S310). at least a portion of the screen information 300 delivered in the first region 261 based on the preset touch type. More specifically, in a state in which the screen information 300 has been delivered in the first region (front display unit 261) (S310), a touch is detected in the second region 262 disposed on the side surface of the body (S320). The touch sensed at 5320 is a preset touch type associated with an acquisition function for acquiring at least a portion of the screen information 300 delivered in the first region 261. Although not shown, in a state in which the screen information 300 has been delivered in the first region (front display unit 261), if a touch is detected in the second region 262, the controller 180 determines whether the detected touch is a touch related to a performing an acquisition function. The touch associated with an execution of an acquisition function corresponds to a preset type of touch, which can be defined in various ways. The touch related to an execution of an acquisition function will be explained in detail below. If the detected touch is a touch related to an execution of an acquisition function as a result of determination, the controller 180 processes the detected touch as an acquisition command to acquire at least one part of the screen information 300 delivered in the first region 261 (S330). In a state in which the screen information 300 has been delivered to the first region 261, if a preset type of touch relative to an acquisition function is detected on the second region 262, the controller 180 can acquire the 300. At least one of a position and a size of a region of the screen information 300 to be acquired ("acquisition region") may be variable depending on a type. of touch applied to the second region 262. The controller 180 may acquire all or part of the screen information 300 according to a touch applied to the second region 262. Full screen information 300 may be a region instead of a status display bar. The status display bar means a region in which status information of the mobile terminal, such as a current time, the remaining amount of a battery and a reception sensitivity, or environment information is displayed. In the present invention, an acquisition function can be performed through a plurality of steps. For example, if a first touch relative to an acquisition function is applied to the second region 262, the controller 180 may enter an acquisition mode. The acquisition mode means a series of conceptualized processes performed in the mobile terminal until the acquisition is made. In the acquisition mode, the controller 180 may perform an acquisition operation in response to an acquisition command applied to the mobile terminal. For example, in the acquisition mode, an acquisition region can be set, changed, stored, and so on.
[0041] As shown in FIG. 4A, in a state in which the screen information 300 has been delivered to the first region 261, if a preset type of touch relative to the execution of an acquisition mode is detected on the second region 262, the controller 180 may enter (execute) an acquisition mode. In the acquisition mode, the controller 180 can set an acquisition region 410, as shown in FIG. 4B, in response to the preset touch type or to another preset touch type different from said preset type. to touch. If the acquisition region 410 has been set in response to a touch applied to the second region 262, the controller 180 acquires screen information included in the acquisition region 410 in response to a control command. additional control. The additional control command can be applied in various ways. For example, the additional control command may be applied via a touch applied to the second region 262, or a touch applied to the first region 261. The control command may also be applied by the intermediate of a voice or a hardware key provided at the terminal body. The controller 180 may acquire screen information included in the acquisition region 410 in response to the additional control command. If the screen information included in the acquisition region 410 has been acquired, the screen information included in the acquisition region 410 is stored in the memory 170 in an image file format or in another format file. Heretofore, it has been explained that the screen information 300 is acquired in response to an additional control command. Nevertheless, the controller 180 may acquire screen information included in the acquisition region on a touch basis to adjust the acquisition region. In one example, if a touch to adjust the acquisition region 410 is detected for a preset time, the controller 180 can acquire screen information included in the acquisition region 410. In another example, if a pre-set type of one-touch tap to set the acquisition region 410 has been detected, the controller 180 can acquire screen information included in the acquisition region 410.
[0042] In the above embodiment, tuning processes of the acquisition region 410 have been explained. Nevertheless, in the present invention, at least a portion of the screen information 300 can be directly performed, in response to a touch applied to the second region 261, without a process for adjusting the acquisition region 410. As mentioned above, the controller may be provided with an acquisition function for acquiring information supplied to the display unit in response to a touch applied to the side display unit. The present invention can thus overcome limitations in the size of the display unit by using the side display unit, the display unit being configured to receive a control command and to output information. Embodiments of the above-mentioned acquisition function will be described in detail hereinafter with reference to FIGS. 3 and 4. A preset type of touch relative to an acquisition function may be applied to a lateral display unit of various ways. In the present specification, a method of performing an acquisition function using a side display unit will be explained with reference to representative embodiments. A first method of touching, among various touching methods which are operable to perform an acquisition function, will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 5A and 5B, 6A and 6B, 7A to 7D, 8A to 8C, 9A to 9F and 10A to 10C are conceptual views for explaining a method of providing an acquisition function, using a touch applied to a side display unit.
[0043] With reference to FIG. 5A, the mobile terminal according to the present invention can execute an acquisition mode on the basis of the detection of touches applied at two points 420a and 420b on a second region 262. If touches are applied at the two points ( the first point 420a and the second point 420b) are simultaneously detected, the controller 180 determines that the detection is a request to acquire at least a portion of screen information 300 delivered to a first region 261. If touches applied to the first point 420a and the second point 420b are detected, in a state in which the screen information 300 has been delivered to the first region 261, the controller 180 performs an acquisition function to acquire at least a portion of the screen information 300. In the present invention, a region in which touches to perform an acquisition function are applied is not limited. e to the second region. In one example, touches applied to the second region 2662 may be replaced by touches applied to a third region 263 in all embodiments relating to an acquisition function. The controller 180 may be configured to recognize an acquisition function request if the first point 420a and the second point 420b are touched. In addition, the controller 180 may set an acquisition region based on the first touched point 420a and the second touched point 420b. More specifically, if touches applied to the first point 420a and second point 420b on the second region 262 are simultaneously detected, the controller 180 adjusts at least a portion of the screen information 300 delivered to the first region 261. as acquisition region 410 based on touchdowns. The controller 180 can determine a position, a size, a zone, etc. of the acquisition region based on at least one of a position of the first point 420a and the second point 420b and a distance between the first point 420a and the second point 420b. In one example, as shown in Fig. 5A (b), a size and position of an acquisition region can be determined by the first point 420a and the second point 420b. The size and position of the acquisition region is determined according to a position of the first point 420a and the second point 420b on the second region 262. More specifically, a vertical length of the acquisition region 410 corresponds to an interval between points on the second region 262 where touches have been detected (e.g., the first point 420a and the second point 420b). A position of the acquisition region 410 on the first region 231 (e.g., a boundary) corresponds to a touch position applied to the second region 262 (e.g., the first point 420a and the second point 420b). A boundary of the acquisition region 410 may be defined by lines extending from the first point 420a and the second point 420b. Lines can be formed to enter the first region 261, and they can be formed in parallel. The lines may extend upward to opposite sides to touched points on the display unit. In this case, the extension lines may be a substantial boundary line delivered to the display unit, or they may be a virtual boundary line that is not delivered to the display unit. The display unit may be divided into three regions by the lines, and the acquisition region 410 may be defined as a central region among the three divided regions (see the oblique region of Figure 5A (b)). . As shown, the acquisition region 410 may have a quadrangular shape. A horizontal length of the acquisition region 410 may be preset without being influenced by touches applied to the second region 262. For example, the horizontal length of the acquisition region 410 may correspond to a horizontal length of the first region 261. As mentioned above, in the present invention, an acquisition region may be defined according to an interval between points on the second region 262 where touches have been detected, or a touch position. Once two points on the second region 262 are touched by the user, the controller 180 may execute an acquisition mode and may adjust an acquisition region. Fig. 5A illustrates a case in which touches applied to the first point 420a and the second point 420b are applied simultaneously. Nevertheless, the present invention is not limited to this. For example, in a case in which a first touch is applied to a point on the second region 262, then a second touch is applied to another point while the first touch is maintained, the controller 180 can execute a mode. acquisition. Thus, even if two touches are not simultaneously applied at two points on the second region 262, the controller 180 may execute an acquisition mode in response to both touch detection. More specifically, in a state in which a first feel is applied to the first point 420a on the second region 262 as shown in Fig. 5B (a), if a second feel applied to the second point 420b is detected as shown on 5B (b), the controller 180 determines that the first touch and the second touch have been detected. simultaneously. Therefore, the controller 180 may specify an acquisition region, as shown in Fig. 5B (c). With reference to Fig. 5A, in a state in which the acquisition region 410 is set, the controller 180 can acquire screen information included in the acquisition region 410, in response to an additional touch applied. at the second region 262. The additional touch is a touch relative to a control command for storing the screen information included in the acquisition region 410 in the memory in an image file format or in another format. file, which can have a preset type.
[0044] More specifically, the additional touch may be a touch subsequent to at least one of the sensed touches on the first point 420a and the second point 420b (e.g., a first touch and a second touch), and may be a preset type of to touch. Alternatively, the additional touch may be a touch that is not consecutive to sensed touches on the first point 420a and the second point 420b (eg, a first touch and a second touch), and may be a preset type to touch. For example, as shown in FIG. 5A (c), a preset type of touch corresponding to the additional touch may be a touch of scrolling consecutive to at least one of the sensed touches on the first point 420a and the second point 420b (for example a first touch and a second touch), the touch of scroll being applied from one to the other of the first touch and the second touch. The additional touch may be a nipping shrinkage resulting from at least one of the sensations detected on the first point 420a and the second point 420b (eg, a first feel and a second feel). In a case where the narrowing nip is applied to the second region 262, the controller 180 may store screen information included in the acquisition region 410 in the memory in a file format. image or in another file format. In a case in which an interval between points at which the narrowing nip is applied is less than a preset length, the controller 180 may acquire screen information included in the acquisition region 410. for example, if the fingers of a user touching the second region 262 move to almost touch each other, the controller 180 can acquire the screen information.
[0045] In the aforementioned configuration, the additional touch is a nipping shrinkage touching at least one of the sensed touches on the first point 420a and the second point 420b (e.g., a first feel and a second feel). Nevertheless, the extra touch can be a nipping nip that is not consecutive to the first touch and the second touch. As mentioned above, in the present invention, in a case in which a first touch and a second touch are applied in sequence, the acquisition of the screen information 300 can be performed. In this case, the first touch is a touch to enter an acquisition mode and to set an acquisition region (i.e. touches applied to the first point 420a and the second point 420b as shown). in Figs. 5A (a) and (b)). The second touch is an additional touch to acquire screen information included in the acquisition region (i.e., a narrowing pinch feel as shown in Fig. 5C (c)).
[0046] In the present invention, at least two steps (for example, a first step of applying a first touch, and a second step of applying a second touch) may be necessary to perform an acquisition operation. The acquisition mode is completed if the screen information acquisition is performed by applying the first touch and the second touch.
[0047] As shown in Fig. 5A, the controller 180 adjusts an acquisition region in response to touches applied to at least two points 420a and 420b on the second region 262. In the present invention, a graphical effect can be implemented so that information on the set acquisition region can be provided to a user. A user may have difficulty recognizing an operation of the mobile terminal unless information is provided in at least one in an audible manner, in a visual and tactile manner. Even if an acquisition region has been adjusted by a touch applied to the second region 262, a user may not accurately recognize a position of the acquisition region. To solve such a problem, in the present invention, information on a tuned acquisition region is provided. This may provide a user with information on a precise position of the acquisition region on the first region 261. Information on a region set as an acquisition region may be provided in a variety of ways. In one example, information about a region set as an acquisition region can be delivered in a visible manner. In another example, the controller 180 may output a graphic object (or a guide image) to an acquisition region. Then, a user can recognize the acquisition region, through the graphical object (or guide image) delivered to the acquisition region. As shown in Fig. 6A (a), once an acquisition region 410 (see Figs. 5A and 5B) is set by touches applied at two points on the second region 262, the control 180 outputs a graphic object 430a to the acquisition region 410. The controller 180 outputs the graphic object 430a to the acquisition region 410 set when two points on the second region 262 are touched. The graphic object 430a may be the same size as the acquisition region 410. The graphic object 430a may be formed to cover the acquisition region 410. Even though the graphic object 430a has covered the acquisition region 410, information delivered to the acquisition region 410 may be exposed to the outside. The information delivered to the acquisition region 410 may be overlapped by the graphic object in a user recognizable form. The graphic object may have transparency. When the semitransparent graphic object is overlapped by the acquisition region 410, information delivered to the acquisition region 410 may be recognizable to a user. A design of the graphic object 430a can be varied in various ways. In the present invention, information on a degree of progress of acquisition can be provided in at least one in an audible manner, in a visible manner and in a tactile manner.
[0048] As shown in FIG. 5A, in a state in which an acquisition region is set, at the detection of a touch relative to a control command for storing screen information included in the acquisition region 410 in the memory in an image file format or in another file format, the controller 180 can change the appearance of the graphic object 430a with the touch. In one example, as mentioned above, in the section of a nip shrinkage touching at least one of the detected touches on the first point 420a and the second point 420b to set an acquisition region (See Figs. 5A and 5B) (e.g., a first touch and a second touch), the controller 180 may change the graphical object 430a with the nip pinch. Alternatively, at the detection of a narrowing nip that is not consecutive to sensed touches on the first point 420a and the second point 420b to adjust an acquisition region (see Figs. 5A and 5B) (For example a first touch and a second touch), the controller 180 can change the graphical object 430a with the nip pinch. In one example, as shown in Figs. 6A (b) and 6A (c), when a nip nip is applied to a second region 262, the controller 180 may change the appearance of graphic objects 430a, 430b and 430c. The appearance of the graphic objects 430a, 430b and 430c can be gradually changed with the nip pinch. The size of the graphical objects 430a, 430b, and 430c may be reduced depending on a degree of shrink nip progress. An output size of the graphic objects 430a, 430b and 430c can be progressively reduced. In a state in which an output size of the graphic objects 430a, 430b and 430c is reduced to a size smaller than a reference size, the controller 180 can acquire screen information included in the acquisition region. 410. As mentioned above, the controller 180 can acquire screen information included in the acquisition region 410, if a gap between two points at which the narrowing pinch taps have been applied is less than one. preset value. In one example, if the fingers of a user in contact with the second region 262 move until they almost touch each other, the controller 180 can acquire the screen information. At the end of the acquisition, the output of the graphic objects 430a, 430b and 430c on the display unit 251 can be terminated. An end time of output of the graphic objects 430a, 430b and 430c corresponds to an acquisition end time. Nevertheless, the end of exit time and the end of acquisition time are not necessarily identical. As mentioned above with reference to Figs. 6A (a), (b) and (c), in a state in which the graphic object 430a is output, even if the appearance of the graphic object 430a is changed to of a touch applied to the second region 262 (e.g., a nip nip), the controller 180 may further deliver another graphical object 432 to provide information on a controlled acquisition region, such as is shown in Figures 6B (b) and (c). The controller 180 can change the graphic objects 430a, 430b and 430c with the touch, and it can further deliver another graphic object 432 indicating a position of the acquisition region 410. Said other graphic object 432 can be issued only on an edge of the acquisition region 410. Alternatively, said other graphic object 432 may be delivered on overlapping portions of the acquisition region 410, i.e. the parts to which the graphic objects 430a 430b and 430c have not been issued. In the present invention, even if the appearance of the graphic objects 430a, 430b and 430c is changed with the touch, the controller 180 can continuously provide information on a position of the acquisition region 410, via said other graphic object 432. In the present invention, an acquisition mode may be deactivated (released or terminated) after the end of acquiring screen information included in an acquisition region. In addition, the acquisition mode may be deactivated (released or terminated) even if screen information included in an acquisition region has not been fully acquired. More specifically, the acquisition mode can be canceled before the end of the acquisition. The cancellation of the acquisition mode can be performed by a user's intention.
[0049] In one example, a control command for canceling an acquisition mode can be applied via a touch applied to the second region 262, or a touch applied to the first region 261. The control command can also be applied through a voice or a hardware key provided at the terminal body. An embodiment for canceling an acquisition mode will be explained in detail below. The cancellation of an acquisition mode can be performed on the basis of a touch applied to the second region 262, a loosening of a touch, etc.
[0050] In one example, as mentioned above with reference to FIGS. 5A and 5B, an acquisition mode is activated when touches applied to the two points 420a and 420b on the second region 262 are detected. As shown in Figs. 7A (a) and (b), the acquisition mode may be canceled when the touches applied to the two points 420a and 420b on the second region 262 are released. In a state in which an acquisition region 430 is set by a first touch and a second touch applied to the two points 420a and 420b on the second region 262, if a graphics object is delivered in the acquisition region 430 as is shown in Figs. 6A and 6B, the graphic object may disappear (it may no longer be delivered) if the touches applied to the two points 420a and 420b are released. In another example, as shown in Figs. 7B (a) and (b), if a first touch and a second touch applied to the two points 420a and 420b on the second region 262 are detected, an acquisition mode can to be executed. Even when the first touch and the second touch applied to the two points 420a and 420b on the second region 262 are released, the acquisition mode may not be canceled. In this case, even if the touches applied to the two points 420a and 420b have been released, the controller 180 can maintain the acquisition mode. In a state in which the acquisition mode is maintained, an output state of a graphic object in the acquisition region 430 can be maintained. As shown in Figs. 7B (c) and (d), the acquisition mode can be released when a preset touch type is applied to a region that is not set as an acquisition region 430 in the first region 261. The preset type of touch can be various. In one example, the preset type of touch can be a short touch, a long touch, a double touch, and so on. As shown in FIGS. 6A and 6B, in a state in which a graphic object is delivered to the acquisition region 430, the graphic object may disappear (it may no longer be delivered) if the acquisition mode is canceled (or completed). In another example, with reference to Figs. 7C (a) and (b) as mentioned above with reference to Fig. 7B, in a state in which an acquisition mode is activated when a first touch and a second touch applied to the two points 420a and 420b on the second region 262 are detected, even if the touches applied to the two points 420a and 420b are released, the acquisition mode can be maintained. This means that the controller 180 can maintain the acquisition mode even if the touches applied to the two points 420a and 420b are released. In a state in which the acquisition mode is maintained, an output state of a graphic object in the acquisition region 430 can be maintained. As shown in Figs. 7C (c) and (d), the acquisition mode can be released when a preset type of touch is applied to the second region 262. The preset touch can be varied. In one example, the preset touch can be a short touch, a long touch, a double touch, and so on. Further, as shown in Figs. 6A and 6B, in a state in which a graphic object is delivered to the acquisition region 430, the graphic object may disappear (it may no longer be delivered) if the mode acquisition is canceled (or completed). In another example, as shown in Figs. 7D (b) and (c), the acquisition mode may be canceled via a function icon 440 supplied to the display unit 251. The function icon 440 is an icon relating to the cancellation of an acquisition mode. Such a function icon 440 may be delivered to the display unit 251 in various cases. If an acquisition mode is enabled, the function icon 440 may be output to the display unit 251. If the acquisition mode is deactivated (released or terminated), the function icon 440 may disappear. of the display unit 251 (or it may no longer be delivered to the display unit 251). The function icon 440 may be delivered to at least one of the first region 261 and the second region 262 of the display unit 251. In one example, in a state in which an acquisition mode is activated when a first touch and a second touch applied to the two points 420a and 420b on the second region 262 are detected as shown in Fig. 7D (a), if the touches applied to the two points 420a and 420b are released, the function icon 440 may be delivered to the second region 262 as shown in Figure 7D (b). The acquisition mode can be maintained even if the touches applied to the two points 420a and 420b are released. In a state in which the function icon 440 is issued, if a function relating to the cancellation of an acquisition mode is selected as shown in FIGS. 7D (c) and (d), the control member 180 can cancel the acquisition mode. In the acquisition mode, an output state of a graphic object at the acquisition region 430 can be maintained. If the acquisition mode is canceled or deactivated, the graphic object may disappear (it may no longer be delivered). In the mobile terminal according to the present invention, if an additional command for acquiring screen information has not been applied for a preset time in an acquisition mode, the acquisition mode can be released. In this case, the acquisition mode can be automatically released even if an additional control command has not been entered by a user. A method of changing (resetting) a region set as an acquisition region, in particular a method of changing a position of an acquisition region, will be explained in detail hereinafter with reference to the drawings. attached. In the mobile terminal according to the present invention, an acquisition mode is activated if touches applied to the two points 420a and 420b on the second region 262 are detected as shown in Figs. 5A and 5B. In a state in which an acquisition region is set (see 410 in Fig. 5A and 435a in Fig. 8A) based on a position of a first feel and a second feel applied to both 420a and 420b, etc., the controller 180 can reset the acquisition region in response to touch detection consecutive to the first touch and the second touch, respectively.
[0051] For example, if a slip entry, which starts at first touch and second touch respectively applied to the first point 420a and the second point 420b (see Figure 8A (a)), is applied to a third point 420c and a fourth point 420d as shown in Fig. 8A (b), the controller 180 can reset the acquisition region in response to the slip input (see 435a in Fig. 8A (b)). The controller 180 may change the acquisition region on the basis of a sliding feel subsequent to the first touch and the second touch applied to the first point 420a and the second point 420b. A position of the acquisition region can therefore be changed based on the sliding feel. The touch of sliding following the first touch and the second touch may be touches moving in the same direction. A degree of change in the position of the acquisition region may be variable depending on a degree of progression of the sliding feel or length of the slip feeler. In the aforementioned configuration, the sliding touch is consecutive touches to touches detected at the first point 420a and the second point 420b, for example a first touch and a second touch. Nevertheless, the sliding feel may be a touch that is not consecutive to the first touch and the second touch. In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. If a plurality of taps starting from any two points in the second region 262 are slid in the same direction, the acquisition region may be reset in correspondence with the direction of slip. Further, in a state in which a graphics object is delivered in the acquisition region 430 as shown in FIGS. 6A and 6B, the graphic object can be moved in correspondence with the sliding touch. The graphic object can be delivered in a reset acquisition region in correspondence with the sliding touch. In another example, the reset of the acquisition region may be performed based on a sliding feel starting at a point in the second region 262.
[0052] The feel applied to said point (see Fig. 420d of Fig. 8B (c)) may be a touch reapplied after releasing the applied seals in the second region 262, as shown in Figs. 8B (a) and (b) . In a case where the touches applied in the second region 262 are released, the controller 180 may change the acquisition region based on a touch reapplied to the display unit 251 after releasing the touches. In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. The touch applied to said one point (see Fig. 420d in Fig. 8B (c)) may be one of the first touch and the second feel applied to the first point 420a and the second point 420b (see Figs. 5A and 5B). . As shown in FIGS. 8B (c) and (d), the controller 180 can change the acquisition region on the basis of a sliding feel 15 starting at a point in the second region 262. change of the acquisition region may correspond to the reset of the acquisition region. Through the change (of the reset), a position of the acquisition region can be modified (see 435a and 435b of Figure 8B). The position of the acquisition region can be changed according to one of a direction and a length of the sliding touch. As shown in Figs. 6A and 6B, in a state in which the graphic object is delivered in the acquisition region 430, the graphic object can be moved in correspondence with a touch to change the acquisition region. The graphic object may be delivered to another acquisition region reinitialized by a touch. In another example, the reset of the acquisition region can be performed based on a sliding feel starting at a point in the first region 261. As shown in FIGS. 8C (a) and (b) the touch applied to said one point may be a touch reapplied to the second region 262 after the release of the touches. In a case where the touches applied to the second region 262 are released, the controller 180 may change the acquisition region based on a touch reapplied to the display unit 251 after releasing the touches. . In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. As shown in FIG. 8C (c), the touch applied to said point can be a touch applied to a preset acquisition region 435a. The touch applied to a preset acquisition region 435a may be a slip feel. As shown in FIGS. 8C (c) and (d), the controller 180 can change the acquisition region on the basis of a sliding feel starting from the acquisition region on the first region 261. The change of the acquisition region may correspond to the reset of the acquisition region. Through the change (reset), a position of the acquisition region may be altered (see 435a and 435b of Figure 8C).
[0053] The position of the acquisition region can be changed according to one of a direction and a length of the sliding touch. As shown in Figs. 6A and 6B, in a state in which a graphics object is delivered in the acquisition region 430, the graphical object can be moved in correspondence with a touch to change the acquisition region. The graphical object may be delivered to another acquisition region reinitialized by a touch. Although not shown, the touch applied in the first region 261 to change the acquisition region is not necessarily limited to a touch applied to a preset acquisition region.
[0054] A method of changing (resetting) a region set as an acquisition region, particularly a method of changing a size of an acquisition region, will be explained in detail hereinafter with reference to the drawings. attached. In the mobile terminal according to the present invention, an acquisition mode is activated when touches applied to the two points 420a and 420b in the second region 262 are detected as shown in FIGS. 5A and 5B. In the acquisition mode, the controller 180 may change a size of an acquisition region adjusted based on a preset touch detected in at least one of the first region and the second region. The controller 180 may change a size of an acquisition region set based on a pre-set touch subsequent to at least one of a first touch and a second touch applied to the first point 420a. and at the second point 420b in the second region 262. Alternatively, the controller 180 may change a size of an acquisition region adjusted based on a preset touch that is not consecutive to a first touch. and a second touch applied to the first point 420a and the second point 420b on the second region 262. The non-consecutive touch can be applied to at least one of the first region 261 and the second region 262. In one example in a state in which an acquisition region is set (see 410 in Fig. 5A and 430a in Fig. 9A) based on a position of a first touch and a second touch applied to the two points 420a and 420b 15 on the second region 26 2, etc., the controller 180 may change a size of the acquisition region in response to touch detection consecutive to at least one of the first touch and the second touch. As shown in FIG. 9A (a), the controller 180 can detect a sliding feel 420c starting from the second touch applied to the second point 420b, and not from the first touch applied to the first point. 420a. As shown in Fig. 9A (b), the controller 180 can reset the acquisition region in response to the sliding feel (see 430d in Fig. 9A (b)). The controller 180 can therefore change a size of the acquisition region on the basis of a sliding feel subsequent to one of the first touch and the second touch respectively applied to the first point 420a and the second point 420b. . A degree of change in the size of the acquisition region may be variable depending on a degree of progression of the sliding feel or slip touch length. Further, in a state in which a graphic object is delivered in the acquisition region 430 as shown in FIGS. 6A and 6B, the graphic object can be moved in one-touch correspondence to change the size of the region of acquisition. The graphical object may be delivered to another acquisition region reinitialized by a touch. In another example, the reset of the acquisition region may be performed based on a preset touch starting from a point in the second region 262. The preset touch may be a slip touch. The touch applied at said point may be a touch reapplied to the display unit after releasing the first feel and the second touch applied to the first point 420a and second point 420b in the second region 262 (see Figs. 8B (a)). and B)). In a case where the touches applied to the second region 262 are released, the controller 180 may change the acquisition region based on a touch reapplied to the display unit 251 after releasing the touches. In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. As shown in Fig. 9B (a), the controller 180 can change the size of the acquisition region based on a sliding feel starting from a point 420c in the second region. 262. A degree of change in the size of the acquisition region may be variable depending on a degree of progression of the sliding feel or length of the slip feel. As shown in Figs. 6A and 6B, in a state in which a graphic object is delivered in the acquisition region 430, the graphical object can be moved in one-touch correspondence to change the acquisition region. The graphic object may be delivered to another acquisition region reset by one touch. In another example, the reset of the acquisition region can be performed on the basis of a preset sliding touch on the second region 262. More specifically, the preset sliding touch can be a sliding touch starting from A touch on the first region 261. The touch applied to said point may be a touch reapplied to the display unit after the first touch and the second touch are applied to the first point 420a and the second point 420b in the second point. region 262 (see Figure 5A (a) and (b)).
[0055] In a case where the touches applied to the second region 262 are released, the controller 180 may change the size of the acquisition region based on a touch reapplied to the display unit 251 after the looseness of the touches. In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. As shown in Fig. 9C (c), the touch applied at said point may be a touch applied to a preset acquisition region 430. The touch applied to the preset acquisition region 430 may be variable. For example, the touch applied to the preset acquisition region 430 may be a slip feel, a nip pinch, an enlargement nip, and so on. More specifically, as shown in Figs. 9C (b) and (c), the controller 180 can change the size of an acquisition region on the first region 261 based on a pinch feel shrinkage applied to the acquisition region. The change in the size of the acquisition region corresponds to the reset of the acquisition region (see 430f in FIG. The size of the acquisition region can be changed depending on the degree of the pinch feel (for example, the length of the touch).
[0056] Further, in a state in which a graphic object is delivered to the acquisition region 430 as shown in Figs. 6A and 6B, the graphic object can be moved in one-touch correspondence to change the acquisition region . The graphic object may be delivered to another acquisition region reinitialized by a touch.
[0057] Although not shown, the touch applied to the first region 261 to change the acquisition region is not necessarily limited to a touch applied to a preset acquisition region. Heretofore, an embodiment for controlling a vertical length of an acquisition region has been explained as a method of controlling the size of the acquisition region. In the present invention, the size of an acquisition region can be controlled by controlling not only a vertical length but also a horizontal length of the acquisition region. The method of controlling a horizontal length of an acquisition region may be performed in the same manner or in a manner similar to that which has been mentioned above with reference to Figs. 9A-9C. A touch to change a horizontal length of an acquisition region may be applied to a region instead of the second region 262, because a control command to control a vertical length of the acquisition region is applied to the second region 262. For example, as shown in Fig. 9D (a), changing a horizontal length of an acquisition region may be performed on a third region 263 instead of the first region 261 and of the second region 262. As shown in Figs. 9D (a) and (b), the controller 180 can decrease a horizontal length of the acquisition region 430, in response to a sliding feel applied in the third region 263 in a first sense. The size of the acquisition region 430 is therefore contracted. As shown in Figs. 9D (b) and (c), the controller 180 may increase a horizontal length of the acquisition region 430, in response to a sliding feel applied to the third region 263 in a second meaning. The size of the acquisition region 430 is thus enlarged. In another example, as shown in Figs. 9E (a) and (b), the size of the acquisition region 430 is contracted in response to a sliding feel applied to the third region 263 in a first sense . In this case, a first side is excluded from the acquisition region 430. More specifically, a boundary of the first side moves to a second side opposite to the first side. As shown in Figs. 9E (b) and (c), the size of the acquisition regions 430 and 431a is contracted in response to a sliding feel applied to the third region 263 in a second direction different from the first direction . In this case, a second side is excluded from acquisition regions 430 and 431a. More specifically, a boundary of the second side moves to the first side opposite the second side. The first side and the second side can be respectively the left side and the right side of the mobile terminal. In another example, as shown in FIGS. 9F (a) and (b), the controller 180 can contract the size of the acquisition region (see 430 and 436), in response to a preset touch applied to an edge of the first region 261.
[0058] The preset touch can be at least one of a slip feel, a nip pinch feel and an enlargement pinch feel. As shown in Figs. 6A and 6B, in a state in which a graphic object is delivered to the acquisition region 430, the graphic object can be moved in one-touch correspondence to change the acquisition region. The graphic object may be delivered to another acquisition region reinitialized by a touch. As mentioned above, in the mobile terminal according to the present invention, a preset acquisition region can be reset based on a touch applied to the display unit. The needs of a user of modification of an acquisition region can thus be satisfied. Heretofore, as shown in Figs. 10A (a) and (b), there has been explained a method of adjusting a portion of the first region 261 as an acquisition region based on touch applied to the two points 420a and 420b on the second region, and then acquisition of information included in the acquisition region. A user may wish to acquire the entire first region 261. To satisfy such user needs, the present invention provides a method of tuning the entirety of the first region 261 as a region of acquisition. the base of a touch applied to the second region 262. For example, as shown in Fig. 10A, the controller 180 can determine whether to adjust the entire first region 261 as a region or to adjust a portion of the first region 261 as an acquisition region based on a distance between the first point 420a and the second point 420b in the second region 262. For example, as shown in Figs. 10A (a) and (b), if a distance "a" between the first point 420a and the second point 420b is less than a preset distance, the control member 180 adjusts a portion of the firstly region 261 as acquisition region 430. A method of adjusting a portion of the first region 261 as an acquisition region has been described above with reference to FIGS. 5A and 5B and its detailed explanation. is therefore omitted below. As shown in Figs. 10A (c) and (d), if a distance "b" between the first point 420a and the second point 420b is greater than a preset distance, the controller 180 can adjust the entire of the first region 261 as the acquisition region 440 and can fully acquire the screen information delivered to the first region 261. If the distance "b" between the first point 420a and the second point 420b is greater than a distance preset, the controller 180 can fully acquire the screen information supplied to the first region 261 without adjusting the acquisition region 440. In another example, the controller 180 can determine whether to adjust the the entire first region 261 as an acquisition region or whether to adjust a portion of the first region 261 as an acquisition region on the basis of a touch applied to the second region 261. More specifically, as shown in FIGS. 5A, 10B (a) and 10B (b), if a narrowing nip is applied to the second region 262, the control member 180 acquires the information of the second region 262. included in a portion of the first region 261. An acquisition execution method in response to a narrowing nip is described above with reference to Figs. 5A and 5B, and its detailed explanation is thus omitted. -after. As shown in Figs. 10B (c) and (d), if an enlargement nip is applied to the second region 262 in a direction opposite to a narrowing nip, the controller 180 acquires the screen information included in the entirety of the first region 261. As mentioned above, even if an acquisition region has been set by touches to switch to an acquisition mode and set a region 25 acquisition (i.e. touches applied to the first point 420a and the second point 420b of Figs. 5A (a) and (b)), if a widening nip is applied to the second region 262, the controller 180 may acquire screen information included in the entirety of the first region 261. In another example, as shown in Fig. 10C (a), the controller 180 may acquire the entire first region 261 in response to touches on the second region 262 and the third region 263. More specifically, the controller 180 may acquire the entire first region 261, in a case where a first feel is detected at one end. of the second region 262 and a second touch is detected at another end of the third region 263. As mentioned above, in the present invention, a specific function or a complete operation of the mobile terminal can be controlled in response to a touch applied to a side display unit. In such a configuration, a user can apply a control command to the mobile terminal without blocking a front display unit. The user can thus more effectively use information delivered to the display unit.
[0059] A method of providing various functions in conjunction with information supplied to the display unit based on a touch applied to a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 11, 12A-12C and 13 are conceptual views for explaining a method of providing various functions instead of an acquisition function, using a touch applied to a side display unit. In the aforementioned embodiment, the acquisition of information delivered to the first region 261 is performed in response to a touch applied to the second region 262. In the present invention, information delivered to the first region 261 can be acquired, or information delivered to the first region 261 may be selected, in response to a different type of touch applied to the second region 262. For example, as shown in Figs. 11 (a) and (b), if a touch narrowing pinch is applied to the second region 262, the controller 180 may acquire screen information included in the acquisition region 430. As shown in Figs. 11 (c) and (d), if an enlargement nip is applied to the second region 262, the controller 180 may output information 460 to executable functions in association with content included in a selected region 430. The functions may relate to modifying content included in the selected region 430 or reusing the content. A method of setting, resetting, and changing an acquisition region is similar or identical to a method of setting, resetting, and changing a selected region that has been described above and whose detailed explanations are therefore omitted below. In the present invention, once an acquisition region 430 is set by the above method with reference to Figs. 5A and 5B, the screen information supplied to the acquisition region 430 can be acquired. In addition to the screen information acquisition function provided to the acquisition region 430, the present invention may provide other executable functions using the screen information supplied to the acquisition region 430. For example, as shown in Fig. 12A (a), if the acquisition region 430 is set in response to a touch applied to the second region 262, the controller 180 may output function icons 445 to the second region 262. The function icons 445 relate to executable functions using screen information supplied to the acquisition region 430. The functions executable using screen information delivered to the acquisition region 430 may be a function relating to the acquisition such as an acquisition function, an acquisition cancel function, a copy function, a cut function and a select function integral, or a modification or reuse function with respect to information included in the acquisition region. In the present invention, as shown in Figs. 12B (a) and (b), not only function icons relating to the acquisition, modification and reuse of information included in an acquisition region, but also function icons associated with applications can be delivered. The applications may be applications having screen information usage functions included in the acquisition region. As shown in Figs. 12A and 12B, such function icons may be outputted to at least one of the first region 261, the second region 262 and the third region 263. The function icons may be issued to different regions according to a type of function, an attribute or a sorting category. In the aforementioned embodiment, screen information delivered to the first region is acquired in response to a touch applied to the second region, regardless of a type of screen information delivered to the first region. If an acquisition-related touch applied to the second region is detected, the controller 180 may acquire at least a portion of the screen information delivered to the first region, regardless of a type of the screen information delivered. to the first region.
[0060] In the present invention, even if an acquisition-related touch applied to the second region is detected, the controller 180 may not perform an acquisition function as a function of screen information delivered to the first region. . In this case, the controller 180 may perform other functions instead of an acquisition function. More specifically, if an acquisition-related touch applied to the second region is detected, the controller 180 can determine whether to recognize touch as a control command to perform an acquisition function or as a control function. that control command to perform another function based on a type of an application corresponding to screen information delivered to the first region.
[0061] For example, in a case in which a specific application is set to perform another function instead of an acquisition function, if screen information delivered to the first region 261 is an application execution screen Specifically, the controller 180 may perform said other function, in response to an acquisition-related touch applied to the second region 262. As shown in FIG. 13 (a), in a state in which information of an application set to not perform a function other than an acquisition function are delivered (for example, screen information of an application related to an Internet function are delivered), if a relative touch at the acquisition applied to the second region 262 is detected, the controller 180 can acquire at least a portion of the screen information delivered to the first region 261. As shown in FIG. FIG. 13 (c), in a state in which screen information of an application relating to a function other than an acquisition function is outputted (e.g. screen information of an application relating to the playback of an animated image are outputted), if an acquisition-related touch applied to the second region 262 is detected, the controller 180 may perform a function other than an acquisition function, such as this is shown in Figure 13 (d). For example, if the screen information is an execution screen of an application relating to the playback of a moving picture, the controller 180 may control a volume in response to touch relative to the acquisition. Further, if the screen information is an execution screen of an application relating to the playback of a moving picture, the controller 180 may perform a function relating to the delivered screen information, such as an enlargement image or contraction of image, in response to touch relative to the acquisition. Those skilled in the art may realize that the various embodiments mentioned above are applicable to embodiments relating to a second type of touch or a third type of touch, as will be explained hereinafter. A second type of touch, among various types of touch that can be used for an acquisition function, will be explained hereinafter with reference to the accompanying drawings. Figs. 14A to 14D, 15A and 15B, and 16A to 16C are conceptual views for explaining another embodiment of a method of providing an acquisition function, using a touch applied to a display unit. lateral. Based on the aforementioned first type of touch, the mobile terminal according to the present invention performs an acquisition function, in response to sensed touches at two points in a second region (or on the side display unit). One method of performing an acquisition function using a side display unit based on another type of touch instead of the first type of touch will be explained hereinafter. The mobile terminal according to the present invention can perform an acquisition mode in response to a touch applied to a point on a side display unit. A second type of touch for performing an acquisition mode in response to a touch applied to a point on a side display unit will be explained in detail below. With reference to FIG. 14A, the controller 180 may perform an acquisition mode in response to the detection of a touch applied at a point 510a in the second region 262. If a touch applied to a point 510a in the second region 262 is detected, the controller 180 determines that the touch is a request to acquire at least a portion of screen information 500 delivered in the first region 261. In a state in which the information of 500 are delivered to the first region 261, if a touch applied to said 510a is detected, the controller 180 performs an acquisition function to acquire at least a portion of the screen information 500. the present invention, a region in which the acquisition-related feel is applied may not be limited to the second region. For example, in all embodiments relating to an acquisition function, a touch applied to the second region 262 may be replaced by a touch applied to the third region 263. If a touch applied to a point 510 in the second region 262 region 262 is detected for more than one preset time, the controller 180 can determine a request to acquire at least a portion of screen information 500 delivered in the first region 261. The touch applied during more than one preset time can be a long touch. In such a configuration, the controller 180 can accurately determine whether a user touch applied to the second region 262 is applied to request an acquisition function. In the present invention, if a touch applied to said one point 510a is detected, the controller 180 can not only recognize the touch as a request to perform an acquisition function, but also set a region of interest. acquisition on the basis of the touch applied in point 510a. More specifically, if a touch applied to a point 510a in the second region 262 is detected, the controller 180 adjusts at least a portion of screen information 300 delivered into the first region 261 as a region of acquisition 530 on the basis of touch. As shown in Figs. 14A (b) and (c), the controller 180 can adjust an acquisition region 530 based on a touch-following slip feel applied at said point 510a. A position, a size, a surface, etc. of the acquisition region 530 are determined on the basis of at least one of a position of a first point 510a and a second point 510b, and a length between the first point 510a and the second point 510a. point 510b. The first point 510a corresponds to said point 510a, and the second point 510b indicates a point where the sliding feel is finished. For example, a size and position of the acquisition region is determined (set) based on a position of the first point 510a and the second point 510b in the second region 262. More specifically, a vertical length of the Acquiring region 530 is an interval (or length) between points in the second region 262 at which touches are detected (e.g., the first point 510a and the second point 510b). A position of the acquisition region 530 in the first region 261 (e.g., a boundary) corresponds to points in the second region 262 to which touches are detected (e.g., the first point 420a and the second point 420b). A boundary of the acquisition region 530 may be defined by lines extending from the first point 510a and the second point 510b. The lines may be formed to penetrate the first region 261 and may be formed in parallel. The lines may extend to opposite sides to touched points on the display unit. In this case, the extension lines may be a substantial boundary line delivered to the display unit, or they may be a virtual boundary line that is not delivered to the display unit. The display unit 20 may be divided into three regions by the lines, and the acquisition region 530 may be defined as a central region among the three divided regions (see the oblique region 530 of Fig. 14A). As shown, the acquisition region 530 can have a quadrangular shape. A horizontal length of the acquisition region 530 may be preset without being influenced by touches applied to the second region 262. For example, the horizontal length of the acquisition region 530 may correspond to a horizontal length of the first region 261. As mentioned above, in the present invention, an acquisition region may be defined according to a gap between two points in the second region 262 where touches are detected, or a touch position. When a user applies a sliding feel from the first point 510a to the second point 510b, the controller 180 can execute an acquisition mode and can set an acquisition region.
[0062] Fig. 5A illustrates an embodiment in which a sliding feel is applied consecutively from a touch applied to the first point 510a (e.g., the first touch). Nevertheless, the present invention is not limited to this. For example, in a state in which an acquisition mode is performed in response to a touch applied to a point in the second region 262, even if the touch applied to said point is relaxed, the acquisition mode can be maintained. In a state in which the acquisition region 530 is set, the controller 180 can acquire screen information included in the acquisition region 530, in response to another touch applied to the unit of acquisition. display. The additional touch is a touch relative to a control command for storing the screen information included in the acquisition region 530 in the memory in an image file format or in another file format, which may have a preset type. More specifically, the additional touch may be a preset touch type applied to at least one of the first region 261 and the second region 262. Alternatively, the additional touch may be a touch that is not consecutive to a touch detected at the first point 510a and the second point 510b, and 20 may be a preset type of touch. For example, as shown in Fig. 14A (d), a pre-adjusted type of touch corresponding to the additional touch may be a touch applied to the acquisition region 530 in the first region 261. The touch applied to the region of acquiring 530 in the first region 261 may be a short touch, a long touch, a double touch, etc. If the pre-adjusted touch type corresponding to the extra touch is applied to the first region 261, the controller 180 can store screen information included in the acquisition region 530 in the memory in an image file format. or in another file format. In another example, as shown in Figs. 14B (a) and (b), the touch preset type corresponding to the additional touch may be a touch applied to the acquisition region 530 in the second region 262. touch applied to the second region 262 may be a short touch, a long touch, a double touch, etc. If the pre-adjusted touch type corresponding to the additional touch is a long touch applied to the second region 262, the touch applied to the second region 262 may be a touch subsequent to a sliding touch applied when adjusting the acquisition region. 530, as shown in Figs. 14A (a), (b) and (c). If a slip feel is maintained at an end point for a preset time, the controller 180 may determine that the touch is an acquisition command for storing screen information included in the acquisition region 530. in the memory. More specifically, in a state in which a sliding feel is applied to the second region 262, if the sliding feel is maintained at an end point for a preset time, the controller 180 can acquire screen information. included in the acquisition region 530. Even if a sliding feel applied to the second region 262 is released, the acquisition mode can be maintained. Even if a sliding feel applied to the second region 262 to execute an acquisition mode and to adjust an acquisition region is released, the controller 180 can maintain the acquisition mode. In another example, the pre-adjusted type of touch corresponding to the additional touch is a nip shrinkage applied to the second region 262. For example, as shown in Figs. 14A (a), 14A (b), 14A (c) and 14C (c), the narrowing pinch feel may be a one-tap narrowing nip (eg, a second feel) detected at the second point 510b (or an end point of a touch of sliding). In a state in which the second touch is maintained, if a touch is applied to another point (e.g., the third point 510d), the control member 180 can detect a nipping nip subsequent to the second touch and at the third touch. If such shrinkage nip is applied, the controller 180 can store screen information included in the acquisition region 530 in the memory in an image file format or in another file format. .
[0063] In a case in which an interval between points at which the narrowing nip is applied is less than a preset length, the controller 180 may acquire screen information included in the acquisition region 530. For example if the fingers of a user in contact with the second region 262 come close to each other so as to almost touch each other, the controller 180 can acquire the screen information. In the above configuration, the additional touch is a nipping shrinkage touching at least one of the touches (e.g., a first touch and a second touch) detected at the first point 510a and the second point 510b. Nevertheless, the extra touch can be a nipping nip that is not consecutive to the first touch and the second touch. As mentioned above, in the present invention, in a case where a first touch and a second touch are applied sequentially, the acquisition of screen information 500 can be performed. In this case, the first touch is a touch to enter an acquisition mode and set an acquisition region; and the second touch is an additional touch to acquire screen information included in the acquisition region. In the present invention, at least two steps (for example, a first step of applying a first touch, and a second step of applying a second touch) may be necessary to perform an acquisition operation. The acquisition mode is completed if the screen information acquisition is performed by applying the first touch and the second touch. In another example, as shown in Fig. 14D, an acquisition operation may be performed via a function icon 540 supplied to the display unit 251. The function icon 540 is an icon relating to an acquisition command or an acquisition cancel command. Such a function icon 540 may be output to the display unit 251 in various cases. If an acquisition mode is activated, the function icon 540 may be outputted to the display unit 251. If the acquisition mode is deactivated, the function icon 540 may disappear from the display unit. 251 display (or may no longer be issued to the display unit 251). The function icon 540 may be outputted to at least one of the first region 261 and the second region 262 of the display unit 251. As in the embodiment mentioned above with reference to FIGS. 6A and 6B, in the present embodiment, a graphic effect can be implemented so that information in a tuned acquisition region can be provided to a user. A user may have difficulty recognizing an operation of the mobile terminal, unless information is delivered in at least one in an audible manner, in a visual and tactile manner. Even if an acquisition region is set by a touch applied to the second region 262, a user may not accurately recognize a position of the acquisition region. To solve such a problem, in the present invention, information on a tuned acquisition region is provided. This may allow a user to have information on a specific position of the acquisition region in the first region 261.
[0064] Information about a region set as an acquisition region can be provided in a variety of ways. For example, information about a region set as an acquisition region can be delivered in a visible manner. The controller 180 may, for example, deliver a graphic object (or a guide image) to an acquisition region. A user can then recognize the acquisition region through the graphic object delivered to the acquisition region. In one example, as shown in Figs. 14A (a) and (b), if a point in the second region 262 (e.g., a first point 510a) is touched, the controller 180 may output an image guide 520 (first graphical object) indicating a start point of an acquisition region. If the acquisition region is set in response to a touch applied to the second region 262, etc., the controller 180 outputs a graphical object 530a (second graphical object) to the acquisition region 530 as shown. in Figure 14A (c).
[0065] The graphic object 530a may be the same size as the acquisition region 530. The graphic object 530a may therefore be formed to cover the acquisition region 530.
[0066] Even if the graphic object 530 covers the acquisition region 530, information delivered to the acquisition region 530 can be exposed to the outside. The information delivered to the acquisition region 530 may be overlapped by the graphic object in a form to be recognizable to a user. The graphic object can be transparent. When the semitransparent graphic object overlaps the acquisition region 530, the information supplied to the acquisition region 530 can be recognizable by a user. A design of the graphic object 530a can be modified in various ways.
[0067] In the present invention, information on a degree of acquisition progress can be provided in at least one in an audible manner, in a visible manner and in a tactile manner. Such graphic objects 520 and 530a may disappear from the display unit (or may no longer be delivered to the display unit) if the acquisition mode is released. At the end of the acquisition of screen information included in the acquisition region, the controller 180 controls the display unit 251 so that the graphic objects 520 and 530a can no longer be delivered. In the present invention, an acquisition mode may be disabled (or released) after the complete acquisition of screen information included in an acquisition region. In addition, the acquisition mode may be disabled (or released) even if screen information included in an acquisition region is not fully acquired. More specifically, the acquisition mode can be canceled before the end of the acquisition. The cancellation of the acquisition mode can be performed voluntarily by a user. For example, a control command for canceling an acquisition mode may be applied via a touch applied to the second region 262, or a touch applied to the first region 261. The control command may also be be applied through a voice or a hardware key provided to the terminal body. An embodiment for canceling an acquisition mode will be explained in detail below. The cancellation of an acquisition mode can be performed on the basis of a touch applied to the second region 262, a loosening of a touch, etc. In one example, as shown in Fig. 14A, an acquisition mode is activated when a touch applied to a point 510a in the second region 262 (see Figs. 14A and 15A) is detected. As shown in Figs. 15 (a) to (d), the acquisition mode may be canceled when the touch applied at said point 510a in the second region 262 (see Figs. 14A and 15A) is released. The acquisition mode can be canceled even after the setting of the acquisition region 530. In a state in which a graphic object is delivered to the acquisition region 530 as shown in FIG. 14A, the graphic object may disappear (or may no longer be delivered) if the touch applied to the second region 262 is released. In another example, after setting an acquisition region in response to a touch applied to the second region 262, even if the touch is released, the acquisition mode can be maintained. The controller 180 can therefore maintain the acquisition mode even if the touch applied to the second region 262 is released. In a state in which the acquisition mode is maintained, an output state of a graphic object in the acquisition region 530 can be maintained. In this case, as shown in Fig. 15B (c), the acquisition mode can be released by a preset touch applied to a region in the first region 261, the region not being adjusted as a region. 530. The preset touch can be various. For example, the preset touch can be a short touch, a long touch, a double touch, and so on. As shown in Fig. 14A, in a state in which a graphic object is delivered to the acquisition region 530, the graphic object may disappear (or may no longer be delivered) if the acquisition mode is canceled. (or finished) In another example, as shown in Fig. 14D, the acquisition mode may be canceled via a function icon 540 supplied to the display unit 251. The function icon 540 can be an icon relating to the cancellation of an acquisition mode. Such a function icon 540 may be output to the display unit 251 in various cases. If an acquisition mode is activated, the function icon 540 may be output to the display unit 251. If the acquisition mode is deactivated, the function icon 540 may disappear from the unit of operation. display 251 (or may no longer be delivered to display unit 251). The function icon 540 may be outputted to at least one of the first region 261 and the second region 262 of the display unit 251. An embodiment relating to an output of the function icon 540 has been mentioned above with reference to FIG. 7D, and its detailed explanations will therefore be omitted hereinafter. An acquisition region set by the aforementioned second type of touch can be moved or changed. More specifically, a method of changing (resetting) a region set as an acquisition region, particularly a method of changing a size or position of an acquisition region, can be performed by an identical or similar way to that of a first type of touch. A method of resetting, moving or changing an acquisition region set by a second touch type will not be explained hereinafter. Detailed explanations of this can be provided with reference to Figs. 8A-8C and 9A-9F. The present invention provides a method of adjusting the entire first region 261 as an acquisition region according to embodiments by a second touch method, in response to the acquisition requirements of the entire region. first region 261 of a user. The entire screen information 300 may be a region instead of a status display bar. The status display bar means a region in which mobile terminal status information, such as a current time, the remaining amount of a battery and reception sensitivity, or environment information is displayed. In one example, as shown in Figs. 16A (a) and (b), if the second region 262 is touched for a preset time, the controller 180 can adjust the entire first region 261 as than acquisition region. More specifically, as shown in Figs. 16A (a) and (b), in a case where a touch applied to the second region 262 is maintained for a preset time without being released, the control member 180 adjusts the entire first region 261 as the acquisition region.
[0068] If a touch is applied to a point 510a in the second region 262 as shown in Figs. 16A (a) and 16B (a), the controller 180 may execute an acquisition mode and may output a picture of 520 (first graphical object) indicating a start point of an acquisition region, as shown in Fig. 16B (b). If the touch applied at said point 510a of the second region 262 is maintained for a preset time without being moved or released, the controller 180 may adjust the entire first region 261 as an acquisition region, as is shown in Figure 16B (c). In this case, a graphic object 535 covering the first region 261 can be delivered. Characteristics of such a graphic object have been mentioned above, and their detailed explanations will therefore be omitted hereinafter. In another example, this is shown in Figs. 16C (a) and (b), in a case in which a sliding feel or a moving feel is applied to the second region 262 with a speed greater than a preset value, the controller 180 can acquire the entire first region 261. The sliding touch or the moving feeler may be a touch subsequent to a touch applied at said point 510a in the second region 262 so as to perform a mode of operation. 'acquisition. The acquisition function of the entire first region 261 can also be performed in response to the input of a sliding touch or a moving feel which is not consecutive to a point 510a in the second region 262. Those skilled in the art may realize that the above embodiments are applicable to the aforementioned embodiments in relation to a first type of touch. A third type of touch, among various types of touch that can be used to perform an acquisition function, will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 17 (a) and 17 (b) are conceptual views for explaining another embodiment of a method of providing an acquisition function using a touch applied to a side display unit. According to the first type of touch and the second type of touch as mentioned above, the mobile terminal enters a mode of acquisition in response to a touch applied to one of the second region 262 and the third region 263.
[0069] A method of enabling a mobile terminal to enter an acquisition mode using two side display units will be explained below. With reference to FIG. 14, the mobile terminal according to the present invention can execute an acquisition mode, in response to the detection of a touch applied to a point 610b in the second region 262 and a touch applied to a point 610a in the third region 263. If the touches applied to the first point 610b and the second point 610a are simultaneously detected, the control member 180 determines that the detection is a request to acquire at least a portion of the information of screen 300 delivered to the first region 261. In a state in which the screen information 300 is delivered to the first region 261, if the touches applied to the first point 610b and the second point 610a are detected, the controller 180 performs an acquisition function for acquiring at least a portion of the screen information 300. In the present invention, the controller 180 can recognize a request for an ac function. quisition in response to touches applied to the first item 610b and the second item 610a. In addition, the controller 180 may set an acquisition region based on the first point 610b and the second point 610a to which touches are applied. More specifically, if touches applied to the first point 610b and the second point 610a are detected simultaneously, the controller 180 adjusts at least a portion of the screen information 300 delivered to the first region 261 as an acquisition region. (see the oblique part of Figure 17 (b)). The controller 180 can determine a position, a size, a surface, etc. the acquisition region based on a position of at least one of the first point 610b and the second point 610a, and a degree of touch progression from the first point 610b and the second point 610a . For example, as shown in Fig. 17 (a), a size and position of an acquisition region may be adjusted based on the first point 610b and the second point 610a. In one example, a size and position of the acquisition region can be determined based on a position of the first point 610b in the second region 262 and a position of the second point 610a in the third region 263.
[0070] It can be preset if you need to determine a size, position, etc. an acquisition region based on the first point 610b or the second point 610a. For example, the controller 180 can determine a size and position of an acquisition region based on a touch applied to the second region 262. On the contrary, the controller 180 can determine a size. and a position of an acquisition region on the basis of a touch applied to the third region 263. As shown in Fig. 17 (b), a vertical length of an acquisition region can be determined on the base of a sliding feel applied to the second region 262 and the third region 263. The sliding touch may be a consecutive or non-consecutive touch at the first point 610b and the second point 610a so as to effect an acquisition mode. A method of performing a third type of touch acquisition mode may be the same as or similar to the aforementioned methods by a first type of touch and a second type of touch (i.e. delivery of a graphic object, a method of moving an acquisition region, a method of changing a size of an acquisition region, a method of resetting an acquisition region, a method of acquiring an entire acquisition region and a method of canceling an acquisition mode). Detailed explanations of the method of executing an acquisition mode by a third type of touch will therefore be omitted hereinafter. In the above embodiments, an acquisition region is set in a state in which screen information is delivered to the display unit, and then the information delivered to the acquisition region is acquired. The present invention also provides a method of setting an acquisition region and then acquiring the screen information supplied to the display unit. The method of setting an acquisition region can be performed in a variety of ways. The method may be executable in one of the aforementioned embodiments by a first to a third type of touch. As shown in Fig. 18 (a), the controller 180 may maintain an acquisition mode until an acquisition mode cancel command is applied to the mobile terminal. In this case, the acquisition mode can be maintained even after the complete acquisition of the screen information included in an acquisition region 710. In this case, the acquisition region 710 is maintained on the first region 261 in a preset state. Screen information delivered to the display unit 251 can be changed by the selection of a user or under the control of the controller 180. An acquisition function can be simultaneously performed with another executable function in the mobile terminal. The acquisition function can be performed in multiple tasks with other functions that are executable in the mobile terminal.
[0071] In a state in which the set acquisition region 710 is maintained, as shown in Figs. 18 (a) and (b), screen information supplied to the display unit 251 can be changed (cf. 700a and 700b). In such a configuration, information included in the acquisition region 710 can be changed. A user can thus acquire a plurality of screen information based on the fact that the set acquisition region 710 is maintained. In a case in which an acquisition mode is maintained, it is possible to solve an inconvenience of a user occurring by applying a control command to execute an acquisition mode so as to acquire other screen information to the end of acquiring screen information. The control command for acquiring information included in the acquisition region 710 can be executed in one of the aforementioned embodiments by a first to a third type of touch, and its detailed explanations will therefore be omitted hereinafter. As shown in Figs. 18 (a) and (b), screen information delivered to the first region 261 can be converted, in response to a preset touch applied to the display unit 251. More specifically, the preset touch can be a sliding touch, a moving feel, etc. Only in a case in which the preset touch is applied to a preset region, the controller 180 may convert the screen information. The preset region may be the acquisition region 710 for example. The screen information to be converted may be a run screen of an application that has been recently executed.
[0072] A method of controlling a camera function using a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 19A and 19B, 20A and 20B, 21A and 21B, 22 and 23A and 23B are conceptual views for explaining a method of controlling a camera function using a touch applied to a side display unit. With reference to Figs. 19A and 19B, the controller 180 may control a camera function based on the detection of touch points applied at two points 810a and 810b in the second region 262. For example, in a state in which a image of a subject acquired by a camera is delivered to the display unit (see Fig. 19A (a)), a specific region can be selected from the preview image in response to a touch applied to a side display unit. More specifically, if touches are applied to the first point 810a and the second point 810b (see Fig. 19A (b)), the controller 180 determines that the touches are a request to select at least a portion of the preview image delivered in the first region 261. The controller 180 may set an acquisition region based on the first point 810a and the second point 810b to which touches are applied. For example, as shown in Fig. 19A (c), if touches applied to the first point 810a and the second point 810b in the second region 262 are detected, the controller 180 adjusts at least a portion of the preview image outputted to the first region 261 as an acquisition region 830 based on the touch. Nevertheless, the present invention is not limited to this. If touches applied at the two points in the second region 262 are detected, the controller 180 adjusts at least a portion of the preview image delivered to the first region 261 as an acquisition region 830. In all the embodiments relating to the acquisition of a photo, a touch applied to the second region may be replaced by a touch applied to a third region. The controller 180 can determine a position, a size, a surface, etc. of the acquisition region based on at least one of a position of the first point 810a and a position of the second point 810b in the second region 262, and a length between the first point 810a and the second point 810b.
[0073] For example, as shown in Fig. 19A (c), a size and position of an acquisition region can be determined by the first point 810a and the second point 810b. The size and position of the acquisition region is determined based on a position of the first point 810a and a position of the second point 810b in the second region 262. More specifically, a vertical length of the acquisition region 830 is an interval between points in the second region 262 at which touches are detected (e.g., the first point 810a and the second point 810b). A position of the acquisition region 830 in the first region 261 (for example, a boundary). corresponds to a touch position applied to the second region 262 (for example, the first point 810a and the second point 810b). A boundary of the acquisition region 830 may be defined by lines extending from the first point 810a and the second point 810b. The lines may be formed to penetrate the first region 261, and they may be formed in parallel. The lines may extend to opposite sides to touched points on the display unit. In this case, the extension lines may be a substantial boundary line delivered to the display unit, or they may be a virtual boundary line that is not delivered to the display unit. The display unit may be divided into three regions by the lines, and the acquisition region 830 may be defined as a central region among the three divided regions (see the oblique region of Fig. 19A (c)). ). As shown, the acquisition region 830 can have a quadrangular shape. A horizontal length of the acquisition region 830 may be preset without being influenced by touches applied to the second region 262. For example, the horizontal length of the acquisition region 830 may correspond to a horizontal length of the first region As noted above, in the present invention, an acquisition region may be defined according to an interval between points in the second region 262 at which touches are detected, or a position of the touches. Nevertheless, the present invention is not limited to this. For example, in a state in which a position, a size, a surface, and so on. of an acquisition region are preset, if touches are applied to the two points 810a and 810b in the second region 262, the acquisition region can be reset.
[0074] Fig. 19A illustrates a case in which touches are applied simultaneously to the first point 810a and the second point 810b. Nevertheless, the present invention is not limited to this. For example, in a state in which a first touch applied to a point in the second region 262 is maintained, even if a second touch is applied to another point in the second region 262, an acquisition region can be set. The controller 180 may adjust an acquisition region if two touches applied to the two points in the second region 262 are detected simultaneously, even if both touching are not applied simultaneously to the two points. In the present invention, a graphic effect can be implemented so that information in a tuned acquisition region can be provided to a user. Information in a region set as an acquisition region can be provided in a variety of ways. For example, information on a region set as an acquisition region can be delivered in a visible manner. In one example, the controller 180 may output a graphic object (or a guide image) to an acquisition region. A user can then recognize the acquisition region through the graphical object (or guide image) delivered to the acquisition region. In one example, once an acquisition region 830 is set by touches applied at two points in the second region 262, the controller 180 can output the graphical object 430a (see FIG. the acquisition region 830. The controller 180 outputs the graphic object 430a to the acquisition region 830 set when two points in the second region 262 are touched. The graphic object 430a may be the same size as the acquisition region 830. The graphic object 430a may be formed to overlie the acquisition region 830. Even if the graphic object 430a covers the acquisition region 830 information delivered to the acquisition region 830 can be exposed to the outside. The information delivered to the acquisition region 830 may be overlapped by the graphic object in a user recognizable form. The graphic object can be transparent. When the semi-transparent graphic object is overlapped by the acquisition region 830, the information delivered to the acquisition region 830 can be recognizable by a user. The graphic object 430a may be modified by the above methods with reference to Figs. 6A (a) to (c). The controller 180 may further provide another graphical object 432 (see Fig. 6B) to provide information on a tuned acquisition region. As a method of issuing another graphic object 432, the above-mentioned method with reference to Figs. 6B (b) and (c) can be applied. In the present invention, the setting of an acquisition region can be terminated (a set acquisition region can be released) after the acquisition of screen information included in the acquisition region (the storage image) is completed. Alternatively, the setting of an acquisition region can be terminated (a set acquisition region can be released) even if the acquisition of screen information included in the acquisition region (image storage ) is not completed. Canceling the setting of an acquisition region can be done by a user's intention. In one example, a control command for canceling the setting of an acquisition region may be applied via a touch applied to the second region 262 or a touch applied to the first region 261. The control control can also be applied via a voice or a hardware key provided to the terminal body. The cancellation of setting of an acquisition region can be changed by the above method with reference to Figs. 7A (a) and (b). In another example, even when the touches applied to the two points 810a and 810b are released, the set acquisition region can be maintained. That is, even if the touches applied to the two points 810a and 810b are released, the controller 180 can maintain the set acquisition region. In this state, a delivery state of the graphic object to the acquisition region can be maintained. As mentioned above with reference to Figs. 7C (c) and (d), the setting of the acquisition region may be relaxed when a preset key is applied to the second region 262. The preset type of touch may be various. For example, the preset type of touch can be a short touch, a long touch, a double touch, and so on.
[0075] In a state in which a graphic object is delivered to the acquisition region, the graphic object may disappear (may no longer be delivered) if the setting of the acquisition region is canceled (or terminated). In another example, as mentioned above with reference to Figs. 7D (b) and (c), the setting of the acquisition region can be performed via the function icon 440 (see Fig. 7D). ) delivered to the display unit 251. In another example, for controlling a camera function based on the touch detection applied to the two points 810a and 810b in the second region 262, a position of a focal point in a preview image can be changed. More specifically, a position of a focal point can be first set based on the first region 261. Then, in a case in which an acquisition region is selected from the preview image by of touches applied to the side display unit, the position of the focal point can be adjusted based on the acquisition region. In one example, with reference to Fig. 19B (a), in a state in which a preview image of a subject acquired by a camera is displayed on the display unit, if touches are applied to the display unit. Lateral display unit, a specific region can be selected from the preview image. The specific region is an acquisition region which may be a region to which a plurality of embodiments mentioned above with reference to Fig. 19A are applied. Referring to Fig. 19B (a), a preview image is output to the first region 261 (or the entire display unit) in a camera mode, and an indicator indicating a focal point is displayed in the first region 261. In the camera mode, a focal point may be formed at a preset position in the first region 261, as shown. For example, a focal point may be formed at a central portion of the first region 261. In this case, if an acquisition region 830 is set by touches applied to the two points 810a and 810b in the second region, the focal point may be formed at a central portion of the acquisition region 830, as shown in Fig. 19B (b). If the preset position is a central part of an acquisition region, a focal point is set to a central portion of a preview image in a preview mode. If an acquisition region is set, a focal point is set to a central part of the acquisition region. An indicator indicating a focal point is first displayed in a central portion of the first region 261. If an acquisition region 830 is set, the indicator is then moved to a central portion of the acquisition region 830. If the acquisition region 830 is released by one of the above methods with reference to Fig. 19A, the position of the focal point can be restored to the home position. If the acquisition region 830 is released, the focal point may be set in a central portion of the first region 261, and the indicator may be delivered to the central portion of the first region 261. Referring to Figs. 19A and 19B in a state in which the acquisition region 830 is set, if an additional control command is applied to the display unit, the controller 180 can store screen information included in the region of the display. 830 acquisition in the form of a photo image. In one example, the control command may be the release of touches applied to the two points 810a and 810b. A control command relating to the storage of screen information included in the acquisition region 830 can be defined in various forms, as will be explained hereinafter with reference to Figs. 20A and 20B. The aforementioned method with reference to Figs. 19A and 19B may be used as a method of setting acquisition region 830. In this embodiment, a control method after adjustment will be explained hereinafter. With reference to Fig. 20A, in a state in which the acquisition region 830 is set, if an additional control command is inputted, it is determined whether to store a preview image or an image delivered to the region. 830 acquisition based on a type of additional control command.
[0076] The additional control command is a touch relative to a control command for storing screen information included in the acquisition region 830 in the memory in an image file format or in another file format, which may have a preset type. More specifically, the additional control command may be a touch subsequent to at least one of the sensed touches on the first point 810a and the second point 810b (for example, a first touch and a second touch), and may be a type preset to touch. Alternatively, the additional control command may be a non-consecutive touch to sensed touches on the first point 810a and the second point 810b (e.g., a first touch and a second touch), and may be a preset touch type. Referring to Fig. 20A (a), in a state in which the acquisition region 830 is set, if a touch is applied to an acquisition button 801, a preview image is stored in the memory as photo file. In this case, the stored image 841 (or an image of the photo file) can be delivered to the display unit. In another example, with reference to Fig. 20A (b), in a state in which the acquisition region 830 is set, if a touch (e.g., a short touch) is applied to the acquisition region 830, the acquisition region 830 in the preview image can be stored as a photo file. The stored image 842 is an image corresponding to the acquisition region 830 in the preview image, which may be outputted to the display unit. Once the image is stored, the setting of the acquisition region 830 can be released.
[0077] In another example, a control command for storing an image corresponding to the acquisition region 830 may be a sliding touch subsequent to at least one of the two-point detected touches (for example, a first touch and a second touch). touch), the sliding feel being applied from one of the first touch and the second touch to each other. A control command for storing an image corresponding to the acquisition region 830 may be a narrowing pinch touch subsequent to at least one of the sensed touches on the first point 810a and the second point 810b (see Fig. 19A ) (for example, a first touch and a second touch). If the aforementioned narrowing pinch feel is applied to the second region 262, the controller 180 may store screen information included in the acquisition region 830 in the memory in an image file format or in another file format. In a case in which a gap between points at which the narrowing nip is applied is less than a preset length, the controller 180 can store screen information included in the acquisition region 830. for example, if the fingers of a user in contact with the second region 262 come close to almost touching each other, the controller 180 may store the screen information.
[0078] In the above configuration, the control command is a narrowing pinch touch subsequent to at least one of a first touch and a second touch. Nevertheless, the control command may be a nipping nip that is not consecutive to the first touch and the second touch. With reference to FIG. 20B, an acquisition button 801 for storing an image delivered to the first region 261 (to which reference is made hereinafter as "first acquisition button"), and an acquisition button 802 for storing an image delivered to the acquisition region 830 (referred to hereinafter as "second acquisition button") can be separately delivered to the display unit. In this case, the second acquisition button 802 may be issued if the acquisition region 830 is set by one of the aforementioned methods. Referring to Fig. 20B (a), in a state in which a first acquisition button 801 and a preview image are issued, if touches are applied to the two points in the second region 262, an acquisition region 830 is set and a second acquisition button 802 is generated. Referring to Fig. 20B (b), in a state in which the acquisition region 830 is set and the second acquisition button 802 is generated, if a touch is applied to the first acquisition button 801, the image Preview is stored in memory as a photo file. In this case, the stored image 841 (or the photo file) can be delivered to the display unit. With reference to Fig. 20B (c), in a state where the acquisition region 830 is set and the second acquisition button 802 is generated, if a touch (for example, a short touch) is applied to the second button 802, the acquisition region 830 in the preview image can be stored as a photo file. The stored image 842 is an image corresponding to the acquisition region 830 in the preview image, which may be outputted to the display unit. Once the image is stored, the setting of the acquisition region 830 can be released.
[0079] An acquisition region set by the aforementioned method with reference to Figs. 19A and 19B may be changed (reset). A method of changing (resetting) an acquisition region, particularly a method of changing a position of an acquisition region, will be explained in detail hereinafter with reference to Figs. 21A and 21B. A method of changing a size of an acquisition region will be explained in detail hereinafter with reference to Fig. 22. In the mobile terminal according to the present invention, in a state in which an acquisition region is set, if an additional control command is applied to the display unit, the controller 180 may reset the acquisition region. The additional control command is a touch relative to a control command to reset the acquisition region, which may have a preset type. More specifically, the additional control command may be a touch subsequent to at least one of the sensed touches on the first point 810a and the second point 810b (for example, a first touch and a second touch), and may be a touch applied to match the preset type. Alternatively, the additional control command may be a touch that is not consecutive to sensed touches on the first point 810a and the second point 810b (for example, a first touch and a second touch), and may be a touch applied to match the preset type. With reference to Figs. 21A and 21B, an acquisition region may be moved by an additional control command not consecutive to a first touch and a second touch. Nevertheless, the present invention is not limited to this. The additional control command may be a control command consecutive to at least one of a first touch and a second touch. More specifically, as shown in Figs. 19A and 19B, if touches applied to the first point 810a and the second point 810b in the second region 262 are detected, an acquisition region 830a is set (see Fig. 21A). at)). In a state in which the acquisition region 830a is set, if a slip feel is applied to the second region 262, the acquisition region 830a is moved in a slip-feel input direction. More specifically, if a touch is applied to a first point 811a in the second region 262 and then the touched point is moved to a second point 811b, the acquisition region 830a is moved in an input direction of the touched point. In this case, the acquisition region 830a is moved while maintaining its size. When detecting touches applied to the first point 810a and the second point 810b in the second region 262, the acquisition region 830a is set to have a preset size. The preset size of the acquisition region 830a is maintained while the acquisition region 830a is moved.
[0080] There may be several points affected. The affected points may be two points in the second region to which touches are applied in order to address Acquisition Area 830a. In this case, the additional control command may be a control command consecutive to at least one of a first touch and a second touch.
[0081] In another example, in a state in which the acquisition region 830a is set, if a slip feel is applied to the acquisition region 830a, the acquisition region 830a is moved in a sense input direction. slip. More specifically, if a touch is applied to a first point 812a in the acquisition region 830a and then the touched point is moved to a second point 812b, the acquisition region 830a is moved in an input direction of the touched point. Referring to Fig. 22, a size of an acquisition region may be changed by an additional control command following a first touch and a second touch. Nevertheless, the present invention is not limited thereto. The additional control command may be a control command not consecutive to at least one of a first touch and a second touch. With reference to Fig. 22 (a), if a first touch and a second touch applied to the first point 810a and the second point 810b in the second region 262 are respectively detected, an acquisition region 830a is set. As shown in Fig. 22 (b), if a sliding feel starting at the first touch and the second touch is applied to a third point 810c and a fourth point 810d, the controller 180 can reset the 830a acquisition region to another 830c acquisition region. A size of the acquisition region can be changed based on the sliding touch. The touch of sliding after the first touch and the second touch can be a touch applied in opposite directions. The degree of change in the size of the acquisition region may be variable depending on a degree of progression of the sliding feel or slip touch length. Once a graphical object is delivered to the acquisition region, the graphical object can be moved in response to the sliding touch. The graphic object can be delivered to the acquisition region 830c reset in response to the sliding touch. In another example, the size of the acquisition region can be changed based on a sliding feel starting at a point in the second region 262.
[0082] The touch applied at said point may be a touch reapplied after the first and second touches applied to the second region 262 are released. In a case where the first touch and the second touch applied to the second region 262 are released, the controller 180 can change the acquisition region based on a touch reapplied to the display unit 251. after the touches are released. In this case, the acquisition mode can be maintained even when at least one of the first touch and the second touch applied to the first point 420a and the second point 420b is released. In another example, the size of the acquisition region may be changed based on a sliding feel starting at a point of the first region 261. The touch applied at said point may be a touch reapplied after the first touch and the second touch applied to the first region 261 are released. In this case, one of an upper edge and a lower edge of the acquisition region can be moved in an input direction of the sliding touch. Since one of the upper edge and the lower edge of the acquisition region is fixed, the size of the acquisition region is changed. The touch applied to a point may be a touch applied to a preset acquisition region. A touch applied to the second region 262 or the acquisition region 30 may have various types. For example, the touch applied to the acquisition region may be a slip feel, a nip pinch feel, a widening nip, and so on.
[0083] More specifically, in the same manner as the method of changing the size of the acquisition region shown in Figs. 9C (b) and (c), the controller 180 may change the size of the acquisition region on the basis of a shrinkage nip applied to an acquisition region in the first region 261. The change in the size of the acquisition region corresponds to the reset of the acquisition region. The size of the acquisition region can be changed depending on the degree of the pinch feel (for example, the length of the touch). After changing the position or size of the acquisition region, a preview image or an image included in the acquisition region may be stored by the above-mentioned method with reference to Figs. 20A and 20B. After setting an acquisition region, another function relating to a camera function, for example a zoom function, can be performed. A method of performing a zoom function will be explained in detail below with reference to Figs. 23A and 23B. In the mobile terminal according to the present invention, in a state in which an acquisition region is set, if an additional control command is applied to the display unit, the controller 180 can perform a zoom function. . With reference to Figs. 23A and 23B, a zoom function may be performed by an additional control command not consecutive to a first touch and a second touch to adjust an acquisition region. Nevertheless, the present invention is not limited to this. The additional control command may be a control command consecutive to at least one of a first touch and a second touch. More specifically, if touches applied to the first point 810a and the second point 810b in the second region 262 (see Fig. 19A) are detected, an acquisition region 830a is set. In this state, if a sliding feel is applied to the second region 262 as shown in Fig. 23A, a zoom function is performed. More specifically, if a two-point sliding feeler 813a and 813b is moved to two other points 813c and 813d, a zoom function can be performed. For example, if the sliding touch starting at both points 813a and 813b is applied in directions away from each other, an in-zoom function can be performed. On the other hand, if the sliding touch starting at the two points 813a and 813b is applied in directions approaching each other, an outgoing zoom function can be performed. The degree of zoom in / out of a preview image may be variable depending on a degree of progression of the slip feel or slip touch length. As shown, in the case of performing an incoming zoom function, a preview image may be fully zoomed in, and an image included in the acquisition region 830a may be zoomed in.
[0084] In the case where a zoom function is performed by a subsequent control command, the two points 813 and 813b at which the sliding touch begins may be the first point 810a and the second point 810b which define the acquisition region 830a. In another example, as shown in Fig. 23B, in a state in which an acquisition region 930 is set, if a touch is applied to the acquisition region 930, a zoom function can be performed. In this case, the touch applied to the acquisition region 930 can have various types. In one example, the touch applied to the acquisition region 930 can be a slip feel, a nip pinch, a widening nip, and so on. As shown, in the case of performing a zoom function, a preview image may be fully zoomed in or out, and an image included in the acquisition region 830a may be zoomed in or out. outgoing. After setting the acquisition region and executing the zoom function, a preview image or an image included in the acquisition region may be stored by the above method with reference to Figs. 20A and 20B. A method of controlling a function relating to a moving image using a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 24A-24C are conceptual views for explaining a method of controlling a function relating to a moving image using a touch applied to a side display unit. Referring to Figs. 24A-24C, a function relating to the reading of a moving image may be controlled, in response to the detection of a touch on at least one point of the second region 262. Nevertheless, the present invention does not is not limited to that. The controller 180 may control a function relating to the playback of a moving image in response to the detection of a touch on at least one point of the third region 263. In all embodiments, a touch applied to the second region can be replaced by a touch applied to the third region. Referring to Fig. 24A (a), in a case where an animated image is read (or the reading is terminated) when a motion picture playback application is controlled, if touches are detected at two points 910a and 910b of the second region 262, the mobile terminal enters a moving picture playback control mode. The control mode indicates a mode for controlling a function relating to the playback of an animated image in response to the input of an additional control command. As shown, in the control mode, the controller 180 may output control icons to control the playback of an animated image on the display unit. The control icons can be a backspace key, a pause key, a fast forward key, a play time indicator, and so on. If a touch (for example, a short touch) is applied to a command icon, a corresponding operation is performed. The controller 180 may further display, on the display unit, a representative image of moving image data (for example, a thumbnail of a motion picture file), information about data. animated images (for example, a title of an animated image), etc. Referring to Fig. 24A (b), if a touch is applied to the second region in the control mode, a function relating to a moving picture can be controlled. In this case, the touch applied to the second region may be a sliding feel consecutive to at least one of the sensations detected at the first point 910a and the second point 910b (for example, a first touch and a second touch). The touch applied to the first point 910a can be maintained at its fixed position. More specifically, after a touch has been applied to the second point 910b in the second region 262, if the touched point is moved to a third point 910c, a playing time of a moving picture is changed. The read time indicator may be moved by a specific section 920 corresponding to a moving distance of the touched point (or a sliding touch distance). An animated image corresponding to a corresponding playing time can be delivered.
[0085] In another example, as shown in Fig. 24A (b), the specific section 920 can be adjusted by sliding input consecutive to at least one of a first touch and a second touch. In this case, a section control mode for entering a control command relating to the specific section 920 can be executed. The section control mode is executed at the same time as the setting of the specific section 920. In the section control mode, repeat, cut, select, and so on. as for the specific section 920 can be activated. In the section control mode, an input window 920a having a repetition icon, a cutoff icon and a selection icon may be outputted to the display unit. The repeat icon, cut icon, and selection icon are respectively related to a section repeat function, a section cut function, and a section selection function. If a user touches the repeat icon, a moving picture is repeated in the specific section 920. If a user touches the cut icon, the moving picture data corresponding to the specific section 920 can be stored in a file. additional. In another example, if a user touches the selection icon, the specific section 920 is selected, and a current state is converted to a state in which an additional control command with respect to the selected specific section 920 (e.g. , marking, etc.) can be entered. In another example, with reference to Fig. 24B, a moving picture playback control mode may be executed in response to the detection of a touch applied to a point in the second region 262. Referring to Fig. 24C, a specific function of the control mode may be effected by an additional touch applied to the second region 262. More specifically, with reference to Fig. 24B (a), in a case where an animated image is read (or the reading is complete) when a motion picture playback application is controlled, if a touch is detected at a point in the second region 262, the mobile terminal enters a moving picture playback control mode. In the control mode, the controller 180 may output control icons to control the playback of an animated image on the display unit. The control icons can be a backspace key, a pause key, a fast forward key, a play time indicator, and so on. If a touch (for example, a short touch) is applied to one of the control icons, a corresponding operation can be performed. Referring to Fig. 24B (b), a progress bar 930 relating to an animated image to be read can be delivered in the second region 262 by a touch to execute a control mode. The second region 262 is formed to be long in one direction and the progress bar 930 is arranged in said direction. A read time indicator may be output to the progress bar 930. With reference to Fig. 24C, if a touch is applied to the second region 262 in the control mode, an animated image playback time may be controlled . In this case, the touch applied to the second region may be a sliding touch subsequent to the feel detected at said point. If a sliding touch is applied to the second region 262 in the control mode, the playback time indicator may move in a sliding-touch input direction, and a moving picture may be played back at a desired time. corresponding reading. More specifically, after applying a touch to a first point 910a in the second region 262, if the touched point is moved to a second point 910b, a playing time of the moving picture is changed. The touched point can be several points. If there are several points hit, a first touch and a second touch are applied to the second region 262 so as to execute a control mode. In this case, a slip input 25 for moving the aforementioned read time may be a control command consecutive to at least one of the first touch and the second touch. Although not shown, in the control mode, a volume control bar instead of the progress bar 930 may be output to the second region 262. An indicator indicating the degree of volume control may be issued to the volume control bar. In this case, if a touch is applied to the second region 262 in the control mode, a volume of a moving picture can be controlled. The touch applied to the second region may be a touch of slip following a touch sensed at a touched point to execute the control mode. In the present invention, if a slip feel is applied to the second region 262 in the control mode. a volume of a moving image can be controlled according to an input direction of the sliding touch.
[0086] A method of controlling functions relating to image output and image modification using a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. Figs. 25 (a) and 25 (b) are conceptual views for explaining a method of controlling an image output function and an image editing function, using a touch applied to an image unit. lateral display. Referring to Fig. 25, an image modification function can be controlled in response to the detection of touches applied to a first point 1010a and a second point 1010b in the second region 262. For example, in a state in which an application relating to a gallery function is performed and a specific image is provided to the display unit, a specific region can be selected in the specific image in response to a touch applied to a side display unit. More specifically, if touches are applied simultaneously to the first point 1010a and the second point 1010b (see Fig. 25 (a)), the controller 180 determines that the touches are a request to select at least a portion an image outputted to the first region 261. The controller 180 may set a modification region based on the first point 1010a and the second point 1010b to which touches have been applied. For example, the controller 180 adjusts at least a portion of an image supplied to the first region 261 as a modification region 1030 based on touches applied to the first point 1010a and the second point 1010b in the second region 262. Nevertheless, the present invention is not limited thereto. If touches applied at two points in the third region 263 are detected, the controller 180 may adjust at least a portion of a preview image delivered to the first region 261 as a modification region 1030 on the base of the touches. In the present embodiment, a touch applied to the second region 262 may be replaced by a touch applied to the third region 263.
[0087] The controller 180 can determine a position, a size, a surface, etc. of the modification region based on at least one of a position of the first point 1010a and the second point 1010b in the second region 262 and a distance between the first point 1010a and the second point 1010b.
[0088] For example, a size and position of a modification region can be determined by the first point 1010a and the second point 1010b. The size and position of the modification region are determined according to a position of the first point 1010a and the second point 1010b in the second region 262. In this case, a vertical length of the modification region 1030 may correspond to a the interval between points in the second region 262 at which touches are detected (e.g., the first point 1010a and the second point 1010b), as in the aforementioned example of an acquisition region. The method of determining a size and position of an acquisition region mentioned above with reference to Figs. 19A and 19B can be applied as a method of determining a size and position of a region of modification. In the present invention, a graphic effect may be implemented so as to provide information on a modification region set to a user. The method of providing a graphical effect in an acquisition region mentioned above with reference to Figs. 19A and 19B may be used as a method of providing a graphical effect in such a modification region. Canceling the setting of a modification region can be performed by a user's intent. For example, a control command for canceling a setting of a modification region can be applied via a touch applied to the second region 262 or a touch applied to the first region 261. The control command can also be applied through a voice or a hardware key provided at the terminal body. The cancellation of the adjustment of a modification region can be changed by the aforementioned method with reference to Figs. 7A (a) and (b).
[0089] As shown in Fig. 25 (b), in a state in which the modification region 1030 is set, if an additional control command is applied to the display unit, the controller 180 can memorize screen information included in modification region 1030 as a photo image. The control command may be a sliding touch subsequent to at least one of the detected two-point touches (for example, a first touch and a second touch), the slip feel being applied from one of the first touch and touch one to the other. A control command for storing an image corresponding to the modification region 1030 may be a narrowing pinch touch subsequent to at least one of the sensed touches at the first point 1010a and the second point 1010b. If the aforementioned narrowing pinch feel is applied to the second region 262, the controller 180 may store screen information included in the modification region 1030 in the memory in an image file format or in another file format. In the above configuration, the control command is a narrowing pinch touch subsequent to at least one of a first touch and a second touch. Nevertheless, the control command may be a nip shrinkage touch not consecutive to at least one of the first touch and the second touch. In the present invention, the additional control control is not limited to a nip nip. For example, the additional control command may be the release of sensed touches at the first point 1010a and the second point 1010b. A method of providing an information retrieval function using a side display unit will be explained in detail hereinafter with reference to the accompanying drawings. Fig. 26 is a conceptual view for explaining a method of controlling a search function, using a touch applied to a side display unit. Referring to Fig. 26, a search function may be performed in response to the detection of touches applied at two points 1110a and 1110b in the second region 262. For example, in a state in which an application having an output function When a list is executed and a list is issued, if a touch is applied to a side display unit, a specific region to search can be selected from the list. More specifically, if touches are applied simultaneously to the first point 1110a and the second point 1110b, the controller 180 determines that the touches are a request to select at least a portion of a list issued to the first region 261. For example, as shown in Fig. 26, in a state in which an application relating to the transmission / reception of message is executed and a list of messages is delivered, if touches are applied to the first point 1110a and the second point 1110b in the second region 262 are detected, the controller 180 sets at least a portion of the messages included in the list issued to the first region 261 as the search region 1130 based on the touch.
[0090] Nevertheless, the present invention is not limited to this. If touches applied at two points in the third region 263 are detected, the controller 180 may adjust at least a portion of the messages included in the list issued to the first region 261 as the search region 1130 based on the touches.
[0091] The controller 180 can determine a position, a size, and so on. from the search region 1130 on the basis of at least one of a position of the first point 1110a and the second point 1110b in the second region 262, and a distance between the first point 1110a and the second point 1110b . The method for determining a size and position of an acquisition region mentioned above with reference to Figs. 19A and 19B can be applied as a method of determining a position and a size of the region of In the present invention, a graphical effect can be implemented so that information about a tuned search region can be provided to a user. The method of providing a graphical effect in an acquisition region mentioned above with reference to Figs. 19A and 19B may be used as a method of providing a graphical effect in such a search region. Canceling the setting of a search region can be done by a user's intent. For example, a control command for canceling the setting of a search region may be applied via a touch applied to the second region 262, or a touch applied to the first region 261. The control control can also be applied via a voice or a hardware key provided at the terminal body. The cancellation of the setting of a search region can be changed by the above method with reference to Figs. 7A (a) and (b). When setting the search region 1130, the mobile terminal can enter a search mode. In this case, an input window 1120 for entering a search keyword can be delivered to the display unit. If touches applied to the first point 1110a and the second point 1110b in the second region 262 are detected, the search region 1130 is set and the input window 1120 is delivered to the display unit. Once the search for a keyword entered in the input window 1120 is executed, messages included in the search region 1130 are searched. In another example, the input window 1120 may be outputted to the display unit by a method different from a method of setting the search region 1130. More specifically, the search region 1130 is set in response. the detection of touches applied to the first point 1110a and the second point 1110b in the second region 262. In this case, an input window exit icon 300 may be displayed on an execution screen of a relative application. the transmission / reception of messages, with a list of messages. In a state in which the search region 1130 is set, if a touch is applied to the input window output icon 300, the input window 1120 for inputting a key word is outputted. In this case, once the search for a keyword entered in the input window 1120 is executed, messages included in the search region 1130 are searched. In another example, the search region 1130 can be adjusted after entering a keyword. More specifically, in a state in which the input window output icon 300 is delivered to an application execution screen (relating to the transmission / reception of messages), with a list of messages, if the icon 300 is touched, the input window 1120 is delivered. Then, if a keyword is entered in the input window 1120, all the search conditions are satisfied. Touching is then applied to the first point 1110a and the second point 1110b in the second region 262 and the search region 1130 is set. A control command for performing a search function may be applied consecutively from touches applied to the first point 1110a and the second point 1110b.
[0092] In one example, the control command for performing a search function may be a sliding touch subsequent to at least one of the detected touches at the first point 1110a and the second point 1110b (for example, a first touch and a second touch ), the sliding touch being applied from one of the first touch and the second touch, towards each other. The additional touch (e.g., a slip feel) may be a nipping shrinkage subsequent to at least one of the sensed touches at the first point 1110a and the second point 1110b (eg, a first touch and a second touch ).
[0093] If the aforementioned narrowing pinch feel is applied to the second region 262, the controller 180 may perform a search function with respect to screen information included in the search region 1130. A control method of mobile terminal in response to a user input applied to a side display unit has been explained so far. Nevertheless, the present invention is not limited to this. The present invention can also be applied to a mobile terminal that does not have a side display unit. For example, a touch sensor may be provided on a side surface of the mobile terminal, instead of the aforementioned side display unit.
[0094] In this case, the controller can provide or control the above-mentioned function in response to a touch input of a user detected by the touch sensor provided on the side surface of the mobile terminal. A touch applied to the side display unit in the aforementioned embodiment may be replaced by a touch applied to a side surface of a mobile terminal having no side display unit. In another example, the present invention may also be applied to a mobile terminal having neither a side display unit nor a side touch sensor. A touch sensor may be arranged for example on a portion (a bevel portion) of the front housing 101 (see Figure 1A) which supports the touch sensor window 151a (see Figure 1A). In this case, the controller 180 may provide or control the aforementioned function in response to a user's touch input applied to the bevel portion on the front surface of the mobile terminal. A touch applied to the side display unit in the aforementioned embodiment may be replaced by a touch applied to a bevel portion of a mobile terminal having the bevel portion on the front surface thereof. As mentioned above, in the present invention, a specific function or operation of the mobile terminal can be controlled in response to a touch applied to a side display unit of the mobile terminal. In such a configuration, a user can apply a control command to the mobile terminal without blocking a front display unit. The user can thus more effectively use information delivered to the display unit. Further, since the front display and the side display unit operate interconnected to one another, a new type of convenience for the user can be ensured. Various embodiments may be implemented using a machine readable medium on which instructions to be executed by a processor are stored to perform various methods described herein. Examples of possible machine readable media include a hard disk (HDD), a solid state disk (SSD), a semiconductor disk (SDD), a read only memory (ROM), a random access memory (RAM) , a CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, the other types of storage media described herein, and combinations thereof. If desired, the machine-readable medium may be embodied as a carrier wave (e.g., a transmission over the Internet). The processor may include the controller 180 of the mobile terminal. The foregoing embodiments and the above-mentioned advantages are purely exemplary and are not intended to limit the present disclosure. The present teachings can be immediately applied to other types of apparatus. This description is intended to be illustrative and should not limit the scope of the appended claims. Many variations, modifications and variations will be apparent to those skilled in the art. The features, structures, methods, and other features of the exemplary embodiments described herein may be combined in various ways to provide exemplary and / or alternative exemplary embodiments.
[0095] Since the present functionalities can be realized in various forms without departing from their characteristics, it is to be understood that the embodiments described above are not limited by details of the foregoing description unless otherwise indicated, but they must be widely considered within the scope of the present disclosure as defined in the appended claims. All changes and modifications within the scope of the appended claims or within limits equivalent to the limits of the appended claims are therefore intended to be encompassed within the scope of the appended claims.
权利要求:
Claims (20)
[0001]
REVENDICATIONS1. A mobile terminal (100, 200), comprising: a body including a front side and a rear side; a display (151, 251) comprising a first region (261) disposed on the front side and a second region (262) extending from the first region (261), wherein the second region (262) forms at least a portion of a first edge of the mobile terminal (100, 200) and the first edge is between the front side and the back side; a touch sensor; and a controller (180) configured to: cause the display (151, 251) to display content on the first region (261), wherein the content is not displayed on the second region (262); acquiring a region of the displayed content in response to receiving at least one touch input to the second region (262) detected by the touch sensor.
[0002]
The mobile terminal (100, 200) of claim 1, wherein a size of the acquired region corresponds to the at least one touch input. 20
[0003]
The mobile terminal (100, 200) of claim 1, wherein: the at least one touch input comprises a first touch input and a second touch input; and the acquired region is defined between a position on the first region (261) corresponding to the first touch input and a position on the first region (261) corresponding to the second touch input.
[0004]
The mobile terminal (100, 200) according to claim 3, wherein a position of the first touch input corresponds to an upper limit of the acquired region and a position of the second touch input corresponds to a lower limit of the acquired region. .0
[0005]
The mobile terminal (100, 200) of claim 1, wherein the at least one touch input comprises a slip input and a size of the acquired region is defined by a length of the slip input.
[0006]
The mobile terminal (100, 200) of claim 5, further comprising a memory (170), wherein the controller (180) is configured to cause the memory (170) to store information associated with the region acquired.
[0007]
The mobile terminal (100, 200) according to claim 5, wherein the controller (180) is further configured to acquire the region when the size of the acquired region is changed to be less than a threshold size. in response to the slip input.
[0008]
The mobile terminal (100, 200) according to claim 7, wherein the controller (180) is further configured to cause the display (151, 251) to display a boundary indicator indicating a boundary of the region to acquire when the slip entry is slid into the second region (262).
[0009]
The mobile terminal (100, 200) according to claim 1, wherein the controller (180) is further configured to cause the display (151, 251) to display an indicator region overlapping the displayed content. and corresponding to the acquired region, wherein the indicator region is transparently displayed so that the overlapped content is visible. 25
[0010]
The mobile terminal (100, 200) according to claim 9, wherein the controller (180) is further configured to cease displaying the indicator region in response to a touch input outside of the indicator region displayed. 30
[0011]
The mobile terminal (100, 200) of claim 9, wherein the controller (180) is further configured to change a height of the acquired region and a height of the displayed indicator region in response to a narrowing or narrowing pinch entry1 in the second region (262).
[0012]
The mobile terminal (100, 200) according to claim 11, wherein: the display (151, 251) further comprises a third region extending from the first region (261) away from the second region (262), wherein the third region forms at least a portion of a second edge of the mobile terminal (100, 200) and the second edge is between the front side and the back side; and the controller (180) is further configured to change a width of the acquired region and a width of the displayed indicator region in response to a narrowing or nip nip third region.
[0013]
The mobile terminal (100, 200) according to claim 1, further comprising a memory (170), wherein the content comprises text information and the controller (180) is further configured to output the display (151, 251) to display a graphical user interface (GUI) screen comprising a plurality of options for storing the acquired region, the plurality of options including storing the acquired region in an image file, or a text file.
[0014]
The mobile terminal (100, 200) of claim 1, further comprising a memory (170), wherein the controller (180) is further configured to: cause the memory (170) to store associated information to the acquired region; Identifying an application configured to perform a specific function using the stored information; and causing the display (151, 251) to display an icon associated with the identified application.
[0015]
The mobile terminal (100, 200) of claim 1, further comprising a memory (170), wherein: the at least one touch input comprises a first touch input and a second consecutive touch input; and the controller (180) is further configured to: 2 activate an acquisition mode in response to the first touch input in the second region (262); causing the memory (170) to store the acquired region in response to the second touch input in the second region (262) while the first touch input is maintained; and disable the acquisition mode when the first touch input is released before the second touch input.
[0016]
The mobile terminal (100, 200) according to claim 1, wherein the controller (180) is further configured to acquire the region in response to the at least one touch input based on a type. application being run on the mobile terminal (100, 200).
[0017]
A method of controlling a mobile terminal (100, 200) comprising a front side and a rear side, the method comprising: displaying a content on a first region (261) of a display (151, 251) of the mobile terminal (100, 200), wherein the content is not displayed on a second region (262) of the display (151, 251), the first region (261) being disposed on the front side and the second region (262) extending from the first region (261), wherein the second region (262) forms at least a portion of an edge of the mobile terminal (100, 200) and the edge is between the front side and the back side; and acquiring a region of the displayed content in response to at least one touch input in the second region (262). 25
[0018]
The method of claim 17, wherein the at least one touch input comprises a first touch input and a second touch input, and a position of the first touch input corresponds to an upper limit of the acquired region and a position of the touch input. second touch input corresponds to a lower limit of the acquired region.
[0019]
The method of claim 17, wherein the at least one touch input comprises a slip input and a size of the acquired region is defined by a length of the slip input.
[0020]
The method of claim 17, further comprising displaying an indicator region overlapping the displayed content and corresponding to the acquired region, wherein the indicator region is transparently displayed so that the overlapped content be visible.
类似技术:
公开号 | 公开日 | 专利标题
FR3021133B1|2019-08-30|MOBILE TERMINAL AND METHOD FOR CONTROLLING THE MOBILE TERMINAL
FR3019665A1|2015-10-09|
FR3021424B1|2019-09-20|MOBILE TERMINAL AND METHOD FOR CONTROLLING THE MOBILE TERMINAL
FR3022368B1|2019-06-21|WATCH-TYPE TERMINAL AND CONTROL METHOD THEREOF
US10154186B2|2018-12-11|Mobile terminal and method for controlling the same
FR3025328B1|2019-07-12|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3031601B1|2019-08-30|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3024786A1|2016-02-12|MOBILE TERMINAL AND METHOD FOR CONTROLLING THE SAME
FR3021766A1|2015-12-04|MOBILE TERMINAL AND METHOD FOR CONTROLLING THE MOBILE TERMINAL
FR3021136A1|2015-11-20|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3022649A1|2015-12-25|
FR3022367A1|2015-12-18|
FR3026201A1|2016-03-25|
FR3043478A1|2017-05-12|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3039674A1|2017-02-03|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3021485A1|2015-11-27|MOBILE DEVICE AND METHOD OF CONTROLLING THE SAME
FR3039673A1|2017-02-03|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
FR3021767A1|2015-12-04|MOBILE TERMINAL AND METHOD FOR CONTROLLING THE SAME
US20160098187A1|2016-04-07|Mobile terminal and control method thereof
FR3021425A1|2015-11-27|
FR3041785A1|2017-03-31|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
US20160054567A1|2016-02-25|Mobile terminal, glasses-type terminal, and mutual interworking method using screens thereof
FR3040221A1|2017-02-24|
FR3046470B1|2019-11-08|MOBILE TERMINAL
FR3042084B1|2019-11-08|MOBILE TERMINAL AND METHOD OF CONTROLLING THE SAME
同族专利:
公开号 | 公开日
CN104978136A|2015-10-14|
US20170017390A1|2017-01-19|
US10365821B2|2019-07-30|
FR3019665B1|2017-12-22|
CN104978136B|2018-11-13|
US9392095B2|2016-07-12|
US20150288795A1|2015-10-08|
EP2927792B1|2020-01-01|
EP2927792A1|2015-10-07|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题

US7656393B2|2005-03-04|2010-02-02|Apple Inc.|Electronic device having display and surrounding touch sensitive bezel for user interface and control|
KR101339483B1|2007-03-22|2013-12-10|엘지전자 주식회사|Method for capturing web-page, and mobile terminal thereof|
KR20090005680A|2007-07-09|2009-01-14|삼성전자주식회사|Method for providing gui to offer a cylinderical menu and multimedia apparatus thereof|
KR101373333B1|2007-07-11|2014-03-10|엘지전자 주식회사|Portable terminal having touch sensing based image photographing function and image photographing method therefor|
KR101506488B1|2008-04-04|2015-03-27|엘지전자 주식회사|Mobile terminal using proximity sensor and control method thereof|
EP2207342B1|2009-01-07|2017-12-06|LG Electronics Inc.|Mobile terminal and camera image control method thereof|
US8565829B2|2009-03-02|2013-10-22|Lg Electronics Inc.|Mobile terminal with detachably coupled sub-device|
KR101688134B1|2009-12-16|2016-12-20|엘지전자 주식회사|Mobile terminal having a side touch input device and method for executingfunctions of thereof|
KR101657545B1|2010-05-11|2016-09-19|엘지전자 주식회사|Mobile terminal and operating method thereof|
KR101060175B1|2010-07-08|2011-08-29|한국과학기술원|Method for controlling touch screen, recording medium for the same, and method for controlling cloud computing|
KR20120084861A|2011-01-21|2012-07-31|삼성전자주식회사|Method for capturing screen in portable terminal|
EP2527968B1|2011-05-24|2017-07-05|LG Electronics Inc.|Mobile terminal|
US20130104032A1|2011-10-19|2013-04-25|Jiyoun Lee|Mobile terminal and method of controlling the same|
WO2012097759A2|2012-03-06|2012-07-26|华为终端有限公司|Touch screen operation method and terminal|
KR101515623B1|2012-05-14|2015-04-28|삼성전자주식회사|Method and apparatus for operating functions of portable terminal having bended display|
CN102902454B|2012-08-10|2016-02-03|广州市动景计算机科技有限公司|Content at edge of webpage system of selection and device and mobile terminal|
KR102044460B1|2012-11-12|2019-11-13|엘지전자 주식회사|Mobile terminal and method for controlling of the same|KR20150098115A|2014-02-19|2015-08-27|엘지전자 주식회사|Mobile terminal and method for controlling the same|
KR102252448B1|2014-09-12|2021-05-14|삼성전자주식회사|Method for controlling and an electronic device thereof|
KR102368044B1|2014-11-03|2022-02-25|삼성전자주식회사|User terminal device and method for controlling the user terminal device thereof|
USD786300S1|2014-12-16|2017-05-09|Bioinventors & Entrepeneurs Network Llc|Display screen or portion thereof with a graphical user interface|
KR20160098752A|2015-02-11|2016-08-19|삼성전자주식회사|Display device and method for display thereof and computer-readable recording medium|
KR102252225B1|2015-02-27|2021-05-14|삼성전자주식회사|A method for managing one or more notifications and an electronic device therefor|
KR20170046977A|2015-10-22|2017-05-04|삼성전자주식회사|Electronic device comprising bended display and method for controlling the same|
KR20170046915A|2015-10-22|2017-05-04|삼성전자주식회사|Apparatus and method for controlling camera thereof|
CN105607859B|2016-03-02|2018-12-11|腾讯科技(深圳)有限公司|A kind of data processing method and device|
KR20170129372A|2016-05-17|2017-11-27|삼성전자주식회사|Electronic device comprising display|
CN112527230A|2019-09-17|2021-03-19|华为技术有限公司|Parameter adjusting method and electronic equipment|
CN112600961A|2019-09-19|2021-04-02|华为技术有限公司|Volume adjusting method and electronic equipment|
法律状态:
2016-03-29| PLFP| Fee payment|Year of fee payment: 2 |
2017-03-27| PLFP| Fee payment|Year of fee payment: 3 |
2017-04-28| PLSC| Publication of the preliminary search report|Effective date: 20170428 |
2018-02-26| PLFP| Fee payment|Year of fee payment: 4 |
2019-02-11| PLFP| Fee payment|Year of fee payment: 5 |
2020-02-13| PLFP| Fee payment|Year of fee payment: 6 |
2021-02-16| PLFP| Fee payment|Year of fee payment: 7 |
2022-02-16| PLFP| Fee payment|Year of fee payment: 8 |
优先权:
申请号 | 申请日 | 专利标题
KR20140040153A|KR101504238B1|2014-04-03|2014-04-03|Mobile terminal and control method for the mobile terminal|
KR1020140089958A|KR20160009416A|2014-07-16|2014-07-16|Mobile terminal and control method for the mobile terminal|
[返回顶部]